Merge branch 'deprecations' into 'develop'

Deprecations

See merge request pleroma/pleroma!4011
This commit is contained in:
feld 2023-12-20 21:13:35 +00:00
commit cbdd134174
33 changed files with 74 additions and 56 deletions

View File

@ -0,0 +1 @@

View File

@ -138,7 +138,7 @@ defp set_postgres_server_version do
num num
else else
e -> e ->
Logger.warn( Logger.warning(
"Could not get the postgres version: #{inspect(e)}.\nSetting the default value of 9.6" "Could not get the postgres version: #{inspect(e)}.\nSetting the default value of 9.6"
) )

View File

@ -34,7 +34,7 @@ defp handle_result({:error, message}), do: raise(VerifyError, message: message)
defp check_welcome_message_config!(:ok) do defp check_welcome_message_config!(:ok) do
if Pleroma.Config.get([:welcome, :email, :enabled], false) and if Pleroma.Config.get([:welcome, :email, :enabled], false) and
not Pleroma.Emails.Mailer.enabled?() do not Pleroma.Emails.Mailer.enabled?() do
Logger.warn(""" Logger.warning("""
To send welcome emails, you need to enable the mailer. To send welcome emails, you need to enable the mailer.
Welcome emails will NOT be sent with the current config. Welcome emails will NOT be sent with the current config.
@ -53,7 +53,7 @@ defp check_welcome_message_config!(result), do: result
def check_confirmation_accounts!(:ok) do def check_confirmation_accounts!(:ok) do
if Pleroma.Config.get([:instance, :account_activation_required]) && if Pleroma.Config.get([:instance, :account_activation_required]) &&
not Pleroma.Emails.Mailer.enabled?() do not Pleroma.Emails.Mailer.enabled?() do
Logger.warn(""" Logger.warning("""
Account activation is required, but the mailer is disabled. Account activation is required, but the mailer is disabled.
Users will NOT be able to confirm their accounts with this config. Users will NOT be able to confirm their accounts with this config.
Either disable account activation or enable the mailer. Either disable account activation or enable the mailer.

View File

@ -24,7 +24,7 @@ def check_exiftool_filter do
filters = Config.get([Pleroma.Upload]) |> Keyword.get(:filters, []) filters = Config.get([Pleroma.Upload]) |> Keyword.get(:filters, [])
if Pleroma.Upload.Filter.Exiftool in filters do if Pleroma.Upload.Filter.Exiftool in filters do
Logger.warn(""" Logger.warning("""
!!!DEPRECATION WARNING!!! !!!DEPRECATION WARNING!!!
Your config is using Exiftool as a filter instead of Exiftool.StripLocation. This should work for now, but you are advised to change to the new configuration to prevent possible issues later: Your config is using Exiftool as a filter instead of Exiftool.StripLocation. This should work for now, but you are advised to change to the new configuration to prevent possible issues later:
@ -63,7 +63,7 @@ def check_simple_policy_tuples do
|> Enum.any?(fn {_, v} -> Enum.any?(v, &is_binary/1) end) |> Enum.any?(fn {_, v} -> Enum.any?(v, &is_binary/1) end)
if has_strings do if has_strings do
Logger.warn(""" Logger.warning("""
!!!DEPRECATION WARNING!!! !!!DEPRECATION WARNING!!!
Your config is using strings in the SimplePolicy configuration instead of tuples. They should work for now, but you are advised to change to the new configuration to prevent possible issues later: Your config is using strings in the SimplePolicy configuration instead of tuples. They should work for now, but you are advised to change to the new configuration to prevent possible issues later:
@ -121,7 +121,7 @@ def check_quarantined_instances_tuples do
has_strings = Config.get([:instance, :quarantined_instances]) |> Enum.any?(&is_binary/1) has_strings = Config.get([:instance, :quarantined_instances]) |> Enum.any?(&is_binary/1)
if has_strings do if has_strings do
Logger.warn(""" Logger.warning("""
!!!DEPRECATION WARNING!!! !!!DEPRECATION WARNING!!!
Your config is using strings in the quarantined_instances configuration instead of tuples. They should work for now, but you are advised to change to the new configuration to prevent possible issues later: Your config is using strings in the quarantined_instances configuration instead of tuples. They should work for now, but you are advised to change to the new configuration to prevent possible issues later:
@ -158,7 +158,7 @@ def check_transparency_exclusions_tuples do
has_strings = Config.get([:mrf, :transparency_exclusions]) |> Enum.any?(&is_binary/1) has_strings = Config.get([:mrf, :transparency_exclusions]) |> Enum.any?(&is_binary/1)
if has_strings do if has_strings do
Logger.warn(""" Logger.warning("""
!!!DEPRECATION WARNING!!! !!!DEPRECATION WARNING!!!
Your config is using strings in the transparency_exclusions configuration instead of tuples. They should work for now, but you are advised to change to the new configuration to prevent possible issues later: Your config is using strings in the transparency_exclusions configuration instead of tuples. They should work for now, but you are advised to change to the new configuration to prevent possible issues later:
@ -193,7 +193,7 @@ def check_transparency_exclusions_tuples do
def check_hellthread_threshold do def check_hellthread_threshold do
if Config.get([:mrf_hellthread, :threshold]) do if Config.get([:mrf_hellthread, :threshold]) do
Logger.warn(""" Logger.warning("""
!!!DEPRECATION WARNING!!! !!!DEPRECATION WARNING!!!
You are using the old configuration mechanism for the hellthread filter. Please check config.md. You are using the old configuration mechanism for the hellthread filter. Please check config.md.
""") """)
@ -274,7 +274,7 @@ def move_namespace_and_warn(config_map, warning_preface) do
if warning == "" do if warning == "" do
:ok :ok
else else
Logger.warn(warning_preface <> warning) Logger.warning(warning_preface <> warning)
:error :error
end end
end end
@ -284,7 +284,7 @@ def check_media_proxy_whitelist_config do
whitelist = Config.get([:media_proxy, :whitelist]) whitelist = Config.get([:media_proxy, :whitelist])
if Enum.any?(whitelist, &(not String.starts_with?(&1, "http"))) do if Enum.any?(whitelist, &(not String.starts_with?(&1, "http"))) do
Logger.warn(""" Logger.warning("""
!!!DEPRECATION WARNING!!! !!!DEPRECATION WARNING!!!
Your config is using old format (only domain) for MediaProxy whitelist option. Setting should work for now, but you are advised to change format to scheme with port to prevent possible issues later. Your config is using old format (only domain) for MediaProxy whitelist option. Setting should work for now, but you are advised to change format to scheme with port to prevent possible issues later.
""") """)
@ -299,7 +299,7 @@ def check_gun_pool_options do
pool_config = Config.get(:connections_pool) pool_config = Config.get(:connections_pool)
if timeout = pool_config[:await_up_timeout] do if timeout = pool_config[:await_up_timeout] do
Logger.warn(""" Logger.warning("""
!!!DEPRECATION WARNING!!! !!!DEPRECATION WARNING!!!
Your config is using old setting `config :pleroma, :connections_pool, await_up_timeout`. Please change to `config :pleroma, :connections_pool, connect_timeout` to ensure compatibility with future releases. Your config is using old setting `config :pleroma, :connections_pool, await_up_timeout`. Please change to `config :pleroma, :connections_pool, connect_timeout` to ensure compatibility with future releases.
""") """)
@ -331,7 +331,7 @@ def check_gun_pool_options do
"\n* `:timeout` options in #{pool_name} pool is now `:recv_timeout`" "\n* `:timeout` options in #{pool_name} pool is now `:recv_timeout`"
end) end)
Logger.warn(Enum.join([warning_preface | pool_warnings])) Logger.warning(Enum.join([warning_preface | pool_warnings]))
Config.put(:pools, updated_config) Config.put(:pools, updated_config)
:error :error

View File

@ -23,7 +23,7 @@ def warn do
You are using old workers in Oban crontab settings, which were removed. You are using old workers in Oban crontab settings, which were removed.
Please, remove setting from crontab in your config file (prod.secret.exs): #{inspect(setting)} Please, remove setting from crontab in your config file (prod.secret.exs): #{inspect(setting)}
""" """
|> Logger.warn() |> Logger.warning()
List.delete(acc, setting) List.delete(acc, setting)
else else

View File

@ -144,7 +144,7 @@ defp update({group, key, value, merged}) do
error_msg = error_msg =
"updating env causes error, group: #{inspect(group)}, key: #{inspect(key)}, value: #{inspect(value)} error: #{inspect(error)}" "updating env causes error, group: #{inspect(group)}, key: #{inspect(key)}, value: #{inspect(value)} error: #{inspect(error)}"
Logger.warn(error_msg) Logger.warning(error_msg)
nil nil
end end
@ -178,12 +178,12 @@ defp restart(started_applications, app, _) do
:ok = Application.start(app) :ok = Application.start(app)
else else
nil -> nil ->
Logger.warn("#{app} is not started.") Logger.warning("#{app} is not started.")
error -> error ->
error error
|> inspect() |> inspect()
|> Logger.warn() |> Logger.warning()
end end
end end

View File

@ -59,7 +59,7 @@ def load do
Logger.info("Found emoji packs: #{Enum.join(packs, ", ")}") Logger.info("Found emoji packs: #{Enum.join(packs, ", ")}")
if not Enum.empty?(files) do if not Enum.empty?(files) do
Logger.warn( Logger.warning(
"Found files in the emoji folder. These will be ignored, please move them to a subdirectory\nFound files: #{Enum.join(files, ", ")}" "Found files in the emoji folder. These will be ignored, please move them to a subdirectory\nFound files: #{Enum.join(files, ", ")}"
) )
end end

View File

@ -56,7 +56,7 @@ defp do_open(uri, %{proxy: {proxy_host, proxy_port}} = opts) do
{:ok, conn, protocol} {:ok, conn, protocol}
else else
error -> error ->
Logger.warn( Logger.warning(
"Opening proxied connection to #{compose_uri_log(uri)} failed with error #{inspect(error)}" "Opening proxied connection to #{compose_uri_log(uri)} failed with error #{inspect(error)}"
) )
@ -90,7 +90,7 @@ defp do_open(uri, %{proxy: {proxy_type, proxy_host, proxy_port}} = opts) do
{:ok, conn, protocol} {:ok, conn, protocol}
else else
error -> error ->
Logger.warn( Logger.warning(
"Opening socks proxied connection to #{compose_uri_log(uri)} failed with error #{inspect(error)}" "Opening socks proxied connection to #{compose_uri_log(uri)} failed with error #{inspect(error)}"
) )
@ -106,7 +106,7 @@ defp do_open(%URI{host: host, port: port} = uri, opts) do
{:ok, conn, protocol} {:ok, conn, protocol}
else else
error -> error ->
Logger.warn( Logger.warning(
"Opening connection to #{compose_uri_log(uri)} failed with error #{inspect(error)}" "Opening connection to #{compose_uri_log(uri)} failed with error #{inspect(error)}"
) )

View File

@ -70,15 +70,15 @@ def parse_proxy(proxy) when is_binary(proxy) do
{:ok, parse_host(host), port} {:ok, parse_host(host), port}
else else
{_, _} -> {_, _} ->
Logger.warn("Parsing port failed #{inspect(proxy)}") Logger.warning("Parsing port failed #{inspect(proxy)}")
{:error, :invalid_proxy_port} {:error, :invalid_proxy_port}
:error -> :error ->
Logger.warn("Parsing port failed #{inspect(proxy)}") Logger.warning("Parsing port failed #{inspect(proxy)}")
{:error, :invalid_proxy_port} {:error, :invalid_proxy_port}
_ -> _ ->
Logger.warn("Parsing proxy failed #{inspect(proxy)}") Logger.warning("Parsing proxy failed #{inspect(proxy)}")
{:error, :invalid_proxy} {:error, :invalid_proxy}
end end
end end
@ -88,7 +88,7 @@ def parse_proxy(proxy) when is_tuple(proxy) do
{:ok, type, parse_host(host), port} {:ok, type, parse_host(host), port}
else else
_ -> _ ->
Logger.warn("Parsing proxy failed #{inspect(proxy)}") Logger.warning("Parsing proxy failed #{inspect(proxy)}")
{:error, :invalid_proxy} {:error, :invalid_proxy}
end end
end end

View File

@ -173,7 +173,7 @@ def get_or_update_favicon(%URI{host: host} = instance_uri) do
end end
rescue rescue
e -> e ->
Logger.warn("Instance.get_or_update_favicon(\"#{host}\") error: #{inspect(e)}") Logger.warning("Instance.get_or_update_favicon(\"#{host}\") error: #{inspect(e)}")
nil nil
end end
@ -201,7 +201,7 @@ defp scrape_favicon(%URI{} = instance_uri) do
end end
rescue rescue
e -> e ->
Logger.warn( Logger.warning(
"Instance.scrape_favicon(\"#{to_string(instance_uri)}\") error: #{inspect(e)}" "Instance.scrape_favicon(\"#{to_string(instance_uri)}\") error: #{inspect(e)}"
) )
@ -284,7 +284,7 @@ defp scrape_metadata(%URI{} = instance_uri) do
end end
rescue rescue
e -> e ->
Logger.warn( Logger.warning(
"Instance.scrape_metadata(\"#{to_string(instance_uri)}\") error: #{inspect(e)}" "Instance.scrape_metadata(\"#{to_string(instance_uri)}\") error: #{inspect(e)}"
) )

View File

@ -20,7 +20,7 @@ def vacuum(args) do
"full" -> "full" ->
Logger.info("Running VACUUM FULL.") Logger.info("Running VACUUM FULL.")
Logger.warn( Logger.warning(
"Re-packing your entire database may take a while and will consume extra disk space during the process." "Re-packing your entire database may take a while and will consume extra disk space during the process."
) )

View File

@ -73,7 +73,7 @@ def handle_continue(:init_state, _state) do
data_migration.state == :manual or data_migration.name in manual_migrations -> data_migration.state == :manual or data_migration.name in manual_migrations ->
message = "Data migration is in manual execution or manual fix mode." message = "Data migration is in manual execution or manual fix mode."
update_status(:manual, message) update_status(:manual, message)
Logger.warn("#{__MODULE__}: #{message}") Logger.warning("#{__MODULE__}: #{message}")
data_migration.state == :complete -> data_migration.state == :complete ->
on_complete(data_migration) on_complete(data_migration)
@ -109,7 +109,7 @@ def handle_info(:perform, state) do
Putting data migration to manual fix mode. Try running `#{__MODULE__}.retry_failed/0`. Putting data migration to manual fix mode. Try running `#{__MODULE__}.retry_failed/0`.
""" """
Logger.warn("#{__MODULE__}: #{message}") Logger.warning("#{__MODULE__}: #{message}")
update_status(:manual, message) update_status(:manual, message)
on_complete(data_migration()) on_complete(data_migration())
@ -125,7 +125,7 @@ def handle_info(:perform, state) do
defp on_complete(data_migration) do defp on_complete(data_migration) do
if data_migration.feature_lock || feature_state() == :disabled do if data_migration.feature_lock || feature_state() == :disabled do
Logger.warn( Logger.warning(
"#{__MODULE__}: migration complete but feature is locked; consider enabling." "#{__MODULE__}: migration complete but feature is locked; consider enabling."
) )

View File

@ -192,7 +192,7 @@ defp response(conn, client, url, status, headers, opts) do
halt(conn) halt(conn)
{:error, error, conn} -> {:error, error, conn} ->
Logger.warn( Logger.warning(
"#{__MODULE__} request to #{url} failed while reading/chunking: #{inspect(error)}" "#{__MODULE__} request to #{url} failed while reading/chunking: #{inspect(error)}"
) )

View File

@ -70,7 +70,7 @@ def handle_event(
%{key: key}, %{key: key},
_ _
) do ) do
Logger.warn(fn -> Logger.warning(fn ->
"Pool worker for #{key}: Client #{inspect(client_pid)} died before releasing the connection with #{inspect(reason)}" "Pool worker for #{key}: Client #{inspect(client_pid)} died before releasing the connection with #{inspect(reason)}"
end) end)
end end

View File

@ -29,7 +29,7 @@ def filter(%Pleroma.Upload{tempfile: file, content_type: "image" <> _} = upload)
{:ok, :filtered, upload} {:ok, :filtered, upload}
rescue rescue
e in ErlangError -> e in ErlangError ->
Logger.warn("#{__MODULE__}: #{inspect(e)}") Logger.warning("#{__MODULE__}: #{inspect(e)}")
{:ok, :noop} {:ok, :noop}
end end
end end
@ -46,7 +46,7 @@ def filter(%Pleroma.Upload{tempfile: file, content_type: "video" <> _} = upload)
{:ok, :filtered, upload} {:ok, :filtered, upload}
rescue rescue
e in ErlangError -> e in ErlangError ->
Logger.warn("#{__MODULE__}: #{inspect(e)}") Logger.warning("#{__MODULE__}: #{inspect(e)}")
{:ok, :noop} {:ok, :noop}
end end
end end

View File

@ -1560,7 +1560,7 @@ def unmute(muter_id, mutee_id) do
unmute(muter, mutee) unmute(muter, mutee)
else else
{who, result} = error -> {who, result} = error ->
Logger.warn( Logger.warning(
"User.unmute/2 failed. #{who}: #{result}, muter_id: #{muter_id}, mutee_id: #{mutee_id}" "User.unmute/2 failed. #{who}: #{result}, muter_id: #{muter_id}, mutee_id: #{mutee_id}"
) )

View File

@ -293,7 +293,7 @@ defp write(query, dir, name, fun, caller_pid) do
acc + 1 acc + 1
else else
{:error, e} -> {:error, e} ->
Logger.warn( Logger.warning(
"Error processing backup item: #{inspect(e)}\n The item is: #{inspect(i)}" "Error processing backup item: #{inspect(e)}\n The item is: #{inspect(i)}"
) )

View File

@ -475,7 +475,7 @@ def update_outbox(
|> json(message) |> json(message)
e -> e ->
Logger.warn(fn -> "AP C2S: #{inspect(e)}" end) Logger.warning(fn -> "AP C2S: #{inspect(e)}" end)
conn conn
|> put_status(:bad_request) |> put_status(:bad_request)

View File

@ -203,7 +203,7 @@ def config_descriptions(policies) do
if Enum.all?(@required_description_keys, &Map.has_key?(description, &1)) do if Enum.all?(@required_description_keys, &Map.has_key?(description, &1)) do
[description | acc] [description | acc]
else else
Logger.warn( Logger.warning(
"#{policy} config description doesn't have one or all required keys #{inspect(@required_description_keys)}" "#{policy} config description doesn't have one or all required keys #{inspect(@required_description_keys)}"
) )

View File

@ -19,7 +19,7 @@ def filter(message) do
try_follow(follower, message) try_follow(follower, message)
else else
nil -> nil ->
Logger.warn( Logger.warning(
"#{__MODULE__} skipped because of missing `:mrf_follow_bot, :follower_nickname` configuration, the :follower_nickname "#{__MODULE__} skipped because of missing `:mrf_follow_bot, :follower_nickname` configuration, the :follower_nickname
account does not exist, or the account is not correctly configured as a bot." account does not exist, or the account is not correctly configured as a bot."
) )

View File

@ -41,7 +41,7 @@ defp steal_emoji({shortcode, url}, emoji_dir_path) do
shortcode shortcode
e -> e ->
Logger.warn("MRF.StealEmojiPolicy: Failed to write to #{file_path}: #{inspect(e)}") Logger.warning("MRF.StealEmojiPolicy: Failed to write to #{file_path}: #{inspect(e)}")
nil nil
end end
else else
@ -53,7 +53,7 @@ defp steal_emoji({shortcode, url}, emoji_dir_path) do
end end
else else
e -> e ->
Logger.warn("MRF.StealEmojiPolicy: Failed to fetch #{url}: #{inspect(e)}") Logger.warning("MRF.StealEmojiPolicy: Failed to fetch #{url}: #{inspect(e)}")
nil nil
end end
end end

View File

@ -156,7 +156,7 @@ def fix_in_reply_to(%{"inReplyTo" => in_reply_to} = object, options)
|> Map.drop(["conversation", "inReplyToAtomUri"]) |> Map.drop(["conversation", "inReplyToAtomUri"])
else else
e -> e ->
Logger.warn("Couldn't fetch #{inspect(in_reply_to_id)}, error: #{inspect(e)}") Logger.warning("Couldn't fetch #{inspect(in_reply_to_id)}, error: #{inspect(e)}")
object object
end end
else else
@ -182,7 +182,7 @@ def fix_quote_url_and_maybe_fetch(object, options \\ []) do
object object
e -> e ->
Logger.warn("Couldn't fetch #{inspect(quote_url)}, error: #{inspect(e)}") Logger.warning("Couldn't fetch #{inspect(quote_url)}, error: #{inspect(e)}")
object object
end end
end end

View File

@ -550,7 +550,7 @@ def remove_mute(user_id, activity_id) do
remove_mute(user, activity) remove_mute(user, activity)
else else
{what, result} = error -> {what, result} = error ->
Logger.warn( Logger.warning(
"CommonAPI.remove_mute/2 failed. #{what}: #{result}, user_id: #{user_id}, activity_id: #{activity_id}" "CommonAPI.remove_mute/2 failed. #{what}: #{result}, user_id: #{user_id}, activity_id: #{activity_id}"
) )

View File

@ -321,13 +321,13 @@ def date_to_asctime(date) when is_binary(date) do
format_asctime(date) format_asctime(date)
else else
_e -> _e ->
Logger.warn("Date #{date} in wrong format, must be ISO 8601") Logger.warning("Date #{date} in wrong format, must be ISO 8601")
"" ""
end end
end end
def date_to_asctime(date) do def date_to_asctime(date) do
Logger.warn("Date #{date} in wrong format, must be ISO 8601") Logger.warning("Date #{date} in wrong format, must be ISO 8601")
"" ""
end end

View File

@ -201,7 +201,7 @@ defp build_csp_param(url) when is_binary(url) do
def warn_if_disabled do def warn_if_disabled do
unless Config.get([:http_security, :enabled]) do unless Config.get([:http_security, :enabled]) do
Logger.warn(" Logger.warning("
.i;;;;i. .i;;;;i.
iYcviii;vXY: iYcviii;vXY:
.YXi .i1c. .YXi .i1c.

View File

@ -89,7 +89,7 @@ def call(conn, plug_opts) do
end end
defp handle_disabled(conn) do defp handle_disabled(conn) do
Logger.warn( Logger.warning(
"Rate limiter disabled due to forwarded IP not being found. Please ensure your reverse proxy is providing the X-Forwarded-For header or disable the RemoteIP plug/rate limiter." "Rate limiter disabled due to forwarded IP not being found. Please ensure your reverse proxy is providing the X-Forwarded-For header or disable the RemoteIP plug/rate limiter."
) )

View File

@ -9,7 +9,7 @@ defmodule Pleroma.Web.Push do
def init do def init do
unless enabled() do unless enabled() do
Logger.warn(""" Logger.warning("""
VAPID key pair is not found. If you wish to enabled web push, please run VAPID key pair is not found. If you wish to enabled web push, please run
mix web_push.gen.keypair mix web_push.gen.keypair

View File

@ -57,7 +57,7 @@ def perform(
end end
def perform(_) do def perform(_) do
Logger.warn("Unknown notification type") Logger.warning("Unknown notification type")
{:error, :unknown_type} {:error, :unknown_type}
end end

View File

@ -75,7 +75,7 @@ defp log_error(url, {:invalid_metadata, data}) do
end end
defp log_error(url, reason) do defp log_error(url, reason) do
Logger.warn(fn -> "Rich media error for #{url}: #{inspect(reason)}" end) Logger.warning(fn -> "Rich media error for #{url}: #{inspect(reason)}" end)
end end
end end

View File

@ -163,7 +163,7 @@ def find_lrdd_template(domain) do
get_template_from_xml(body) get_template_from_xml(body)
else else
error -> error ->
Logger.warn("Can't find LRDD template in #{inspect(meta_url)}: #{inspect(error)}") Logger.warning("Can't find LRDD template in #{inspect(meta_url)}: #{inspect(error)}")
{:error, :lrdd_not_found} {:error, :lrdd_not_found}
end end
end end

View File

@ -127,7 +127,10 @@ def eval_quoted(block, binding) do
defp start_slave({node_host, override_configs}) do defp start_slave({node_host, override_configs}) do
log(node_host, "booting federated VM") log(node_host, "booting federated VM")
{:ok, node} = :slave.start(~c"127.0.0.1", node_name(node_host), vm_args())
{:ok, node} =
do_start_slave(%{host: "127.0.0.1", name: node_name(node_host), args: vm_args()})
add_code_paths(node) add_code_paths(node)
load_apps_and_transfer_configuration(node, override_configs) load_apps_and_transfer_configuration(node, override_configs)
ensure_apps_started(node) ensure_apps_started(node)
@ -219,4 +222,18 @@ defp node_name(node_host) do
|> Enum.at(0) |> Enum.at(0)
|> String.to_atom() |> String.to_atom()
end end
if System.otp_release() >= "25" do
@peer :peer
else
@peer :slave
end
defp do_start_slave(%{host: host, name: name, args: args} = opts) do
if System.otp_release() >= "25" do
@peer.start(opts)
else
@peer.start(host, name, args)
end
end
end end

View File

@ -41,7 +41,7 @@ defmacro clear_config(config_path, temp_setting) do
# NOTE: `clear_config([section, key], value)` != `clear_config([section], key: value)` (!) # NOTE: `clear_config([section, key], value)` != `clear_config([section], key: value)` (!)
# Displaying a warning to prevent unintentional clearing of all but one keys in section # Displaying a warning to prevent unintentional clearing of all but one keys in section
if Keyword.keyword?(temp_setting) and length(temp_setting) == 1 do if Keyword.keyword?(temp_setting) and length(temp_setting) == 1 do
Logger.warn( Logger.warning(
"Please change `clear_config([section], key: value)` to `clear_config([section, key], value)`" "Please change `clear_config([section], key: value)` to `clear_config([section, key], value)`"
) )
end end

View File

@ -21,7 +21,7 @@ def request(
else else
error -> error ->
with {:error, message} <- error do with {:error, message} <- error do
Logger.warn(to_string(message)) Logger.warning(to_string(message))
end end
{_, _r} = error {_, _r} = error