honkoma/lib/pleroma/formatter.ex

290 lines
7.6 KiB
Elixir
Raw Normal View History

2017-05-17 16:00:09 +00:00
defmodule Pleroma.Formatter do
alias Pleroma.User
alias Pleroma.Web.MediaProxy
2017-05-17 16:00:09 +00:00
@tag_regex ~r/\#\w+/u
def parse_tags(text, data \\ %{}) do
2017-05-17 16:00:09 +00:00
Regex.scan(@tag_regex, text)
2018-03-30 13:01:53 +00:00
|> Enum.map(fn ["#" <> tag = full_tag] -> {full_tag, String.downcase(tag)} end)
|> (fn map ->
if data["sensitive"] in [true, "True", "true", "1"],
do: [{"#nsfw", "nsfw"}] ++ map,
else: map
end).()
2017-05-17 16:00:09 +00:00
end
def parse_mentions(text) do
# Modified from https://www.w3.org/TR/html5/forms.html#valid-e-mail-address
2018-03-30 13:01:53 +00:00
regex =
~r/@[a-zA-Z0-9.!#$%&'*+\/=?^_`{|}~-]*@?[a-zA-Z0-9](?:[a-zA-Z0-9-]{0,61}[a-zA-Z0-9])?(?:\.[a-zA-Z0-9](?:[a-zA-Z0-9-]{0,61}[a-zA-Z0-9])?)*/u
Regex.scan(regex, text)
2018-03-30 13:01:53 +00:00
|> List.flatten()
|> Enum.uniq()
|> Enum.map(fn "@" <> match = full_match ->
{full_match, User.get_cached_by_nickname(match)}
end)
|> Enum.filter(fn {_match, user} -> user end)
end
2017-06-19 15:51:43 +00:00
2017-09-16 13:47:45 +00:00
@finmoji [
"a_trusted_friend",
"alandislands",
"association",
"auroraborealis",
"baby_in_a_box",
"bear",
"black_gold",
"christmasparty",
"crosscountryskiing",
"cupofcoffee",
"education",
"fashionista_finns",
"finnishlove",
"flag",
"forest",
"four_seasons_of_bbq",
"girlpower",
"handshake",
"happiness",
"headbanger",
"icebreaker",
"iceman",
"joulutorttu",
"kaamos",
"kalsarikannit_f",
"kalsarikannit_m",
"karjalanpiirakka",
"kicksled",
"kokko",
"lavatanssit",
"losthopes_f",
"losthopes_m",
"mattinykanen",
"meanwhileinfinland",
"moominmamma",
"nordicfamily",
"out_of_office",
"peacemaker",
"perkele",
"pesapallo",
"polarbear",
"pusa_hispida_saimensis",
"reindeer",
"sami",
"sauna_f",
"sauna_m",
"sauna_whisk",
"sisu",
"stuck",
"suomimainittu",
"superfood",
"swan",
"the_cap",
"the_conductor",
"the_king",
"the_voice",
"theoriginalsanta",
"tomoffinland",
"torillatavataan",
"unbreakable",
"waiting",
"white_nights",
"woollysocks"
]
2018-03-30 13:01:53 +00:00
@finmoji_with_filenames Enum.map(@finmoji, fn finmoji ->
{finmoji, "/finmoji/128px/#{finmoji}-128.png"}
end)
2017-06-19 15:51:43 +00:00
2018-01-31 14:23:31 +00:00
@emoji_from_file (with {:ok, default} <- File.read("config/emoji.txt") do
custom =
with {:ok, custom} <- File.read("config/custom_emoji.txt") do
custom
else
_e -> ""
end
2018-03-30 13:01:53 +00:00
2018-01-31 14:23:31 +00:00
(default <> "\n" <> custom)
|> String.trim()
|> String.split(~r/\n+/)
2018-03-30 13:01:53 +00:00
|> Enum.map(fn line ->
2018-01-31 14:23:31 +00:00
[name, file] = String.split(line, ~r/,\s*/)
{name, file}
2018-03-30 13:01:53 +00:00
end)
2017-09-16 14:49:46 +00:00
else
_ -> []
2018-03-30 13:01:53 +00:00
end)
2017-09-16 14:49:46 +00:00
@emoji_from_globs (
static_path = Path.join(:code.priv_dir(:pleroma), "static")
globs =
Application.get_env(:pleroma, :emoji, [])
|> Keyword.get(:shortcode_globs, [])
paths =
Enum.map(globs, fn glob ->
Path.join(static_path, glob)
|> Path.wildcard()
end)
|> Enum.concat()
Enum.map(paths, fn path ->
shortcode = Path.basename(path, Path.extname(path))
external_path = Path.join("/", Path.relative_to(path, static_path))
{shortcode, external_path}
end)
)
@emoji @finmoji_with_filenames ++ @emoji_from_globs ++ @emoji_from_file
2017-09-16 13:47:45 +00:00
def emojify(text, emoji \\ @emoji)
def emojify(text, nil), do: text
2018-03-30 13:01:53 +00:00
def emojify(text, emoji) do
2018-03-30 13:01:53 +00:00
Enum.reduce(emoji, text, fn {emoji, file}, text ->
2018-03-23 18:52:08 +00:00
emoji = HtmlSanitizeEx.strip_tags(emoji)
file = HtmlSanitizeEx.strip_tags(file)
2018-03-30 13:01:53 +00:00
String.replace(
text,
":#{emoji}:",
"<img height='32px' width='32px' alt='#{emoji}' title='#{emoji}' src='#{
MediaProxy.url(file)
}' />"
)
2018-09-02 00:04:09 +00:00
|> HtmlSanitizeEx.basic_html()
2017-06-19 15:51:43 +00:00
end)
end
2017-09-16 14:14:23 +00:00
def get_emoji(text) do
2018-03-30 13:01:53 +00:00
Enum.filter(@emoji, fn {emoji, _} -> String.contains?(text, ":#{emoji}:") end)
2017-09-16 14:14:23 +00:00
end
2017-10-19 19:51:56 +00:00
def get_custom_emoji() do
@emoji
end
2018-03-24 10:03:10 +00:00
@link_regex ~r/[0-9a-z+\-\.]+:[0-9a-z$-_.+!*'(),]+/ui
# IANA got a list https://www.iana.org/assignments/uri-schemes/ but
# Stuff like ipfs isnt in it
# There is very niche stuff
@uri_schemes [
"https://",
"http://",
"dat://",
"dweb://",
"gopher://",
"ipfs://",
"ipns://",
"irc:",
"ircs:",
"magnet:",
"mailto:",
"mumble:",
"ssb://",
"xmpp:"
]
2018-03-24 10:03:10 +00:00
# TODO: make it use something other than @link_regex
def html_escape(text) do
2018-03-24 10:03:10 +00:00
Regex.split(@link_regex, text, include_captures: true)
|> Enum.map_every(2, fn chunk ->
{:safe, part} = Phoenix.HTML.html_escape(chunk)
part
end)
|> Enum.join("")
end
@doc "changes scheme:... urls to html links"
2018-03-24 10:03:10 +00:00
def add_links({subs, text}) do
additionnal_schemes =
Application.get_env(:pleroma, :uri_schemes, [])
|> Keyword.get(:additionnal_schemes, [])
2018-03-30 13:01:53 +00:00
links =
text
|> String.split([" ", "\t", "<br>"])
|> Enum.filter(fn word -> String.starts_with?(word, @uri_schemes ++ additionnal_schemes) end)
|> Enum.filter(fn word -> Regex.match?(@link_regex, word) end)
|> Enum.map(fn url -> {Ecto.UUID.generate(), url} end)
|> Enum.sort_by(fn {_, url} -> -String.length(url) end)
2018-03-24 10:03:10 +00:00
2018-03-30 13:01:53 +00:00
uuid_text =
links
|> Enum.reduce(text, fn {uuid, url}, acc -> String.replace(acc, url, uuid) end)
2018-03-24 10:03:10 +00:00
2018-03-30 13:01:53 +00:00
subs =
subs ++
Enum.map(links, fn {uuid, url} ->
{:safe, link} = Phoenix.HTML.Link.link(url, to: url)
2018-04-22 08:01:10 +00:00
link =
link
|> IO.iodata_to_binary()
{uuid, link}
2018-03-30 13:01:53 +00:00
end)
2018-03-24 10:03:10 +00:00
{subs, uuid_text}
end
@doc "Adds the links to mentioned users"
def add_user_links({subs, text}, mentions) do
2018-03-30 13:01:53 +00:00
mentions =
mentions
|> Enum.sort_by(fn {name, _} -> -String.length(name) end)
|> Enum.map(fn {name, user} -> {name, user, Ecto.UUID.generate()} end)
uuid_text =
mentions
|> Enum.reduce(text, fn {match, _user, uuid}, text ->
String.replace(text, match, uuid)
end)
subs =
subs ++
2018-04-03 16:32:46 +00:00
Enum.map(mentions, fn {match, %User{ap_id: ap_id, info: info}, uuid} ->
ap_id = info["source_data"]["url"] || ap_id
2018-03-30 13:01:53 +00:00
short_match = String.split(match, "@") |> tl() |> hd()
2018-05-27 13:06:38 +00:00
{uuid,
"<span><a class='mention' href='#{ap_id}'>@<span>#{short_match}</span></a></span>"}
2018-03-30 13:01:53 +00:00
end)
2018-03-24 10:03:10 +00:00
{subs, uuid_text}
end
@doc "Adds the hashtag links"
def add_hashtag_links({subs, text}, tags) do
2018-03-30 13:01:53 +00:00
tags =
tags
|> Enum.sort_by(fn {name, _} -> -String.length(name) end)
|> Enum.map(fn {name, short} -> {name, short, Ecto.UUID.generate()} end)
uuid_text =
tags
|> Enum.reduce(text, fn {match, _short, uuid}, text ->
String.replace(text, match, uuid)
end)
subs =
subs ++
Enum.map(tags, fn {tag_text, tag, uuid} ->
url = "<a href='#{Pleroma.Web.base_url()}/tag/#{tag}' rel='tag'>#{tag_text}</a>"
2018-03-30 13:01:53 +00:00
{uuid, url}
end)
2018-03-24 10:03:10 +00:00
{subs, uuid_text}
end
def finalize({subs, text}) do
2018-03-30 13:01:53 +00:00
Enum.reduce(subs, text, fn {uuid, replacement}, result_text ->
2018-03-24 10:03:10 +00:00
String.replace(result_text, uuid, replacement)
end)
end
2017-05-17 16:00:09 +00:00
end