import Mix.Pleroma
import Ecto.Query
+ import Pleroma.Search.Meilisearch, only: [meili_post!: 2, meili_delete!: 1, meili_get!: 1]
+
def run(["index"]) do
start_pleroma()
- endpoint = Pleroma.Config.get([Pleroma.Search.Meilisearch, :url])
+ meili_post!(
+ "/indexes/objects/settings/ranking-rules",
+ [
+ "desc(published)",
+ "typo",
+ "words",
+ "proximity",
+ "attribute",
+ "wordsPosition",
+ "exactness"
+ ]
+ )
- {:ok, _} =
- Pleroma.HTTP.post(
- "#{endpoint}/indexes/objects/settings/ranking-rules",
- Jason.encode!([
- "desc(published)",
- "typo",
- "words",
- "proximity",
- "attribute",
- "wordsPosition",
- "exactness"
- ])
- )
+ meili_post!(
+ "/indexes/objects/settings/searchable-attributes",
+ [
+ "content"
+ ]
+ )
- chunk_size = 100_000
+ chunk_size = 10_000
Pleroma.Repo.transaction(
fn ->
- Pleroma.Repo.stream(
+ query =
from(Pleroma.Object,
# Only index public posts which are notes and have some text
where:
fragment("data->>'type' = 'Note'") and
- fragment("LENGTH(data->>'source') > 0") and
+ fragment("LENGTH(data->>'content') > 0") and
fragment("data->'to' \\? ?", ^Pleroma.Constants.as_public()),
order_by: [desc: fragment("data->'published'")]
- ),
+ )
+
+ count = query |> Pleroma.Repo.aggregate(:count, :data)
+ IO.puts("Entries to index: #{count}")
+
+ Pleroma.Repo.stream(
+ query,
timeout: :infinity
)
+ |> Stream.map(&Pleroma.Search.Meilisearch.object_to_search_data/1)
+ |> Stream.filter(fn o -> not is_nil(o) end)
|> Stream.chunk_every(chunk_size)
|> Stream.transform(0, fn objects, acc ->
new_acc = acc + Enum.count(objects)
- IO.puts("Indexed #{new_acc} entries")
+ # Reset to the beginning of the line and rewrite it
+ IO.write("\r")
+ IO.write("Indexed #{new_acc} entries")
{[objects], new_acc}
end)
- |> Stream.map(fn objects ->
- Enum.map(objects, fn object ->
- data = object.data
-
- {:ok, published, _} = DateTime.from_iso8601(data["published"])
-
- %{
- id: object.id,
- source: data["source"],
- ap: data["id"],
- published: published |> DateTime.to_unix()
- }
- end)
- end)
|> Stream.each(fn objects ->
- {:ok, result} =
- Pleroma.HTTP.post(
- "#{endpoint}/indexes/objects/documents",
- Jason.encode!(objects)
+ result =
+ meili_post!(
+ "/indexes/objects/documents",
+ objects
)
- if not Map.has_key?(Jason.decode!(result.body), "updateId") do
- IO.puts("Failed to index: #{result}")
+ if not Map.has_key?(result, "updateId") do
+ IO.puts("Failed to index: #{inspect(result)}")
end
end)
|> Stream.run()
end,
timeout: :infinity
)
+
+ IO.write("\n")
end
def run(["clear"]) do
start_pleroma()
+ meili_delete!("/indexes/objects/documents")
+ end
+
+ def run(["show-private-key", master_key]) do
+ start_pleroma()
+
endpoint = Pleroma.Config.get([Pleroma.Search.Meilisearch, :url])
{:ok, result} =
- Pleroma.HTTP.request(:delete, "#{endpoint}/indexes/objects/documents", "", [], [])
+ Pleroma.HTTP.get(
+ Path.join(endpoint, "/keys"),
+ [{"X-Meili-API-Key", master_key}]
+ )
- if not Map.has_key?(Jason.decode!(result.body), "updateId") do
- IO.puts("Failed to clear: #{result}")
+ decoded = Jason.decode!(result.body)
+
+ if decoded["private"] do
+ IO.puts(decoded["private"])
+ else
+ IO.puts("Error fetching the key, check the master key is correct: #{inspect(decoded)}")
end
end
+
+ def run(["stats"]) do
+ start_pleroma()
+
+ result = meili_get!("/indexes/objects/stats")
+ IO.puts("Number of entries: #{result["numberOfDocuments"]}")
+ IO.puts("Indexing? #{result["isIndexing"]}")
+ end
end