+# Pleroma: A lightweight social networking server
+# Copyright © 2017-2019 Pleroma Authors <https://pleroma.social/>
+# SPDX-License-Identifier: AGPL-3.0-only
+
defmodule Pleroma.Formatter do
+ alias Pleroma.Emoji
+ alias Pleroma.HTML
alias Pleroma.User
alias Pleroma.Web.MediaProxy
- @tag_regex ~r/\#\w+/u
- def parse_tags(text, data \\ %{}) do
- Regex.scan(@tag_regex, text)
- |> Enum.map(fn ["#" <> tag = full_tag] -> {full_tag, String.downcase(tag)} end)
- |> (fn map ->
- if data["sensitive"] in [true, "True", "true", "1"],
- do: [{"#nsfw", "nsfw"}] ++ map,
- else: map
- end).()
+ @safe_mention_regex ~r/^(\s*(?<mentions>@.+?\s+)+)(?<rest>.*)/
+ @markdown_characters_regex ~r/(`|\*|_|{|}|[|]|\(|\)|#|\+|-|\.|!)/
+ @link_regex ~r{((?:http(s)?:\/\/)?[\w.-]+(?:\.[\w\.-]+)+[\w\-\._~%:/?#[\]@!\$&'\(\)\*\+,;=.]+)|[0-9a-z+\-\.]+:[0-9a-z$-_.+!*'(),]+}ui
+ # credo:disable-for-previous-line Credo.Check.Readability.MaxLineLength
+
+ @auto_linker_config hashtag: true,
+ hashtag_handler: &Pleroma.Formatter.hashtag_handler/4,
+ mention: true,
+ mention_handler: &Pleroma.Formatter.mention_handler/4
+
+ def mention_handler("@" <> nickname, buffer, opts, acc) do
+ case User.get_cached_by_nickname(nickname) do
+ %User{id: id} = user ->
+ ap_id = get_ap_id(user)
+ nickname_text = get_nickname_text(nickname, opts) |> maybe_escape(opts)
+
+ link =
+ "<span class='h-card'><a data-user='#{id}' class='u-url mention' href='#{ap_id}'>@<span>#{
+ nickname_text
+ }</span></a></span>"
+
+ {link, %{acc | mentions: MapSet.put(acc.mentions, {"@" <> nickname, user})}}
+
+ _ ->
+ {buffer, acc}
+ end
end
- def parse_mentions(text) do
- # Modified from https://www.w3.org/TR/html5/forms.html#valid-e-mail-address
- regex =
- ~r/@[a-zA-Z0-9.!#$%&'*+\/=?^_`{|}~-]+@?[a-zA-Z0-9](?:[a-zA-Z0-9-]{0,61}[a-zA-Z0-9])?(?:\.[a-zA-Z0-9](?:[a-zA-Z0-9-]{0,61}[a-zA-Z0-9])?)*/u
+ def hashtag_handler("#" <> tag = tag_text, _buffer, _opts, acc) do
+ tag = String.downcase(tag)
+ url = "#{Pleroma.Web.base_url()}/tag/#{tag}"
+ link = "<a class='hashtag' data-tag='#{tag}' href='#{url}' rel='tag'>#{tag_text}</a>"
- Regex.scan(regex, text)
- |> List.flatten()
- |> Enum.uniq()
- |> Enum.map(fn "@" <> match = full_match ->
- {full_match, User.get_cached_by_nickname(match)}
- end)
- |> Enum.filter(fn {_match, user} -> user end)
+ {link, %{acc | tags: MapSet.put(acc.tags, {tag_text, tag})}}
+ end
+
+ @doc """
+ Parses a text and replace plain text links with HTML. Returns a tuple with a result text, mentions, and hashtags.
+
+ If the 'safe_mention' option is given, only consecutive mentions at the start the post are actually mentioned.
+ """
+ @spec linkify(String.t(), keyword()) ::
+ {String.t(), [{String.t(), User.t()}], [{String.t(), String.t()}]}
+ def linkify(text, options \\ []) do
+ options = options ++ @auto_linker_config
+
+ if options[:safe_mention] && Regex.named_captures(@safe_mention_regex, text) do
+ %{"mentions" => mentions, "rest" => rest} = Regex.named_captures(@safe_mention_regex, text)
+ acc = %{mentions: MapSet.new(), tags: MapSet.new()}
+
+ {text_mentions, %{mentions: mentions}} = AutoLinker.link_map(mentions, acc, options)
+ {text_rest, %{tags: tags}} = AutoLinker.link_map(rest, acc, options)
+
+ {text_mentions <> text_rest, MapSet.to_list(mentions), MapSet.to_list(tags)}
+ else
+ acc = %{mentions: MapSet.new(), tags: MapSet.new()}
+ {text, %{mentions: mentions, tags: tags}} = AutoLinker.link_map(text, acc, options)
+
+ {text, MapSet.to_list(mentions), MapSet.to_list(tags)}
+ end
+ end
+
+ def emojify(text) do
+ emojify(text, Emoji.get_all())
end
- @finmoji [
- "a_trusted_friend",
- "alandislands",
- "association",
- "auroraborealis",
- "baby_in_a_box",
- "bear",
- "black_gold",
- "christmasparty",
- "crosscountryskiing",
- "cupofcoffee",
- "education",
- "fashionista_finns",
- "finnishlove",
- "flag",
- "forest",
- "four_seasons_of_bbq",
- "girlpower",
- "handshake",
- "happiness",
- "headbanger",
- "icebreaker",
- "iceman",
- "joulutorttu",
- "kaamos",
- "kalsarikannit_f",
- "kalsarikannit_m",
- "karjalanpiirakka",
- "kicksled",
- "kokko",
- "lavatanssit",
- "losthopes_f",
- "losthopes_m",
- "mattinykanen",
- "meanwhileinfinland",
- "moominmamma",
- "nordicfamily",
- "out_of_office",
- "peacemaker",
- "perkele",
- "pesapallo",
- "polarbear",
- "pusa_hispida_saimensis",
- "reindeer",
- "sami",
- "sauna_f",
- "sauna_m",
- "sauna_whisk",
- "sisu",
- "stuck",
- "suomimainittu",
- "superfood",
- "swan",
- "the_cap",
- "the_conductor",
- "the_king",
- "the_voice",
- "theoriginalsanta",
- "tomoffinland",
- "torillatavataan",
- "unbreakable",
- "waiting",
- "white_nights",
- "woollysocks"
- ]
-
- @finmoji_with_filenames Enum.map(@finmoji, fn finmoji ->
- {finmoji, "/finmoji/128px/#{finmoji}-128.png"}
- end)
-
- @emoji_from_file (with {:ok, default} <- File.read("config/emoji.txt") do
- custom =
- with {:ok, custom} <- File.read("config/custom_emoji.txt") do
- custom
- else
- _e -> ""
- end
-
- (default <> "\n" <> custom)
- |> String.trim()
- |> String.split(~r/\n+/)
- |> Enum.map(fn line ->
- [name, file] = String.split(line, ~r/,\s*/)
- {name, file}
- end)
- else
- _ -> []
- end)
-
- @emoji @finmoji_with_filenames ++ @emoji_from_file
-
- def emojify(text, emoji \\ @emoji)
def emojify(text, nil), do: text
- def emojify(text, emoji) do
+ def emojify(text, emoji, strip \\ false) do
Enum.reduce(emoji, text, fn {emoji, file}, text ->
- emoji = HtmlSanitizeEx.strip_tags(emoji)
- file = HtmlSanitizeEx.strip_tags(file)
-
- String.replace(
- text,
- ":#{emoji}:",
- "<img height='32px' width='32px' alt='#{emoji}' title='#{emoji}' src='#{
- MediaProxy.url(file)
- }' />"
- )
+ emoji = HTML.strip_tags(emoji)
+ file = HTML.strip_tags(file)
+
+ html =
+ if not strip do
+ "<img height='32px' width='32px' alt='#{emoji}' title='#{emoji}' src='#{
+ MediaProxy.url(file)
+ }' />"
+ else
+ ""
+ end
+
+ String.replace(text, ":#{emoji}:", html) |> HTML.filter_tags()
end)
end
- def get_emoji(text) do
- Enum.filter(@emoji, fn {emoji, _} -> String.contains?(text, ":#{emoji}:") end)
+ def demojify(text) do
+ emojify(text, Emoji.get_all(), true)
end
- def get_custom_emoji() do
- @emoji
+ def demojify(text, nil), do: text
+
+ def get_emoji(text) when is_binary(text) do
+ Enum.filter(Emoji.get_all(), fn {emoji, _} -> String.contains?(text, ":#{emoji}:") end)
end
- @link_regex ~r/https?:\/\/[\w\.\/?=\-#\+%&@~'\(\):]+[\w\/]/u
+ def get_emoji(_), do: []
+
+ def html_escape({text, mentions, hashtags}, type) do
+ {html_escape(text, type), mentions, hashtags}
+ end
- def html_escape(text) do
+ def html_escape(text, "text/html") do
+ HTML.filter_tags(text)
+ end
+
+ def html_escape(text, "text/plain") do
Regex.split(@link_regex, text, include_captures: true)
|> Enum.map_every(2, fn chunk ->
{:safe, part} = Phoenix.HTML.html_escape(chunk)
|> Enum.join("")
end
- @doc "changes http:... links to html links"
- def add_links({subs, text}) do
- links =
- Regex.scan(@link_regex, text)
- |> Enum.map(fn [url] -> {Ecto.UUID.generate(), url} end)
- |> Enum.sort_by(fn {_, url} -> -String.length(url) end)
-
- uuid_text =
- links
- |> Enum.reduce(text, fn {uuid, url}, acc -> String.replace(acc, url, uuid) end)
+ def truncate(text, max_length \\ 200, omission \\ "...") do
+ # Remove trailing whitespace
+ text = Regex.replace(~r/([^ \t\r\n])([ \t]+$)/u, text, "\\g{1}")
- subs =
- subs ++
- Enum.map(links, fn {uuid, url} ->
- {:safe, link} = Phoenix.HTML.Link.link(url, to: url)
-
- link =
- link
- |> IO.iodata_to_binary()
-
- {uuid, link}
- end)
-
- {subs, uuid_text}
+ if String.length(text) < max_length do
+ text
+ else
+ length_with_omission = max_length - String.length(omission)
+ String.slice(text, 0, length_with_omission) <> omission
+ end
end
- @doc "Adds the links to mentioned users"
- def add_user_links({subs, text}, mentions) do
- mentions =
- mentions
- |> Enum.sort_by(fn {name, _} -> -String.length(name) end)
- |> Enum.map(fn {name, user} -> {name, user, Ecto.UUID.generate()} end)
-
- uuid_text =
- mentions
- |> Enum.reduce(text, fn {match, _user, uuid}, text ->
- String.replace(text, match, uuid)
- end)
-
- subs =
- subs ++
- Enum.map(mentions, fn {match, %User{ap_id: ap_id, info: info}, uuid} ->
- ap_id = info["source_data"]["url"] || ap_id
-
- short_match = String.split(match, "@") |> tl() |> hd()
- {uuid, "<span><a href='#{ap_id}'>@<span>#{short_match}</span></a></span>"}
- end)
-
- {subs, uuid_text}
- end
+ defp get_ap_id(%User{info: %{source_data: %{"url" => url}}}) when is_binary(url), do: url
+ defp get_ap_id(%User{ap_id: ap_id}), do: ap_id
- @doc "Adds the hashtag links"
- def add_hashtag_links({subs, text}, tags) do
- tags =
- tags
- |> Enum.sort_by(fn {name, _} -> -String.length(name) end)
- |> Enum.map(fn {name, short} -> {name, short, Ecto.UUID.generate()} end)
-
- uuid_text =
- tags
- |> Enum.reduce(text, fn {match, _short, uuid}, text ->
- String.replace(text, match, uuid)
- end)
-
- subs =
- subs ++
- Enum.map(tags, fn {_, tag, uuid} ->
- url = "<a href='#{Pleroma.Web.base_url()}/tag/#{tag}' rel='tag'>##{tag}</a>"
- {uuid, url}
- end)
-
- {subs, uuid_text}
- end
+ defp get_nickname_text(nickname, %{mentions_format: :full}), do: User.full_nickname(nickname)
+ defp get_nickname_text(nickname, _), do: User.local_nickname(nickname)
- def finalize({subs, text}) do
- Enum.reduce(subs, text, fn {uuid, replacement}, result_text ->
- String.replace(result_text, uuid, replacement)
- end)
+ defp maybe_escape(str, %{mentions_escape: true}) do
+ String.replace(str, @markdown_characters_regex, "\\\\\\1")
end
+
+ defp maybe_escape(str, _), do: str
end