Refactor connection channel for delivering notifications

This commit is contained in:
Omar Roth 2019-06-03 13:12:06 -05:00
parent 84b2583973
commit d892ba6aa5
No known key found for this signature in database
GPG Key ID: B8254FB7EC3D37F2
2 changed files with 117 additions and 104 deletions

View File

@ -186,10 +186,21 @@ spawn do
end end
end end
notification_channels = [] of Channel(PQ::Notification) connection_channel = Channel({Bool, Channel(PQ::Notification)}).new
PG.connect_listen(PG_URL, "notifications") do |event| spawn do
notification_channels.each do |channel| connections = [] of Channel(PQ::Notification)
channel.send(event)
PG.connect_listen(PG_URL, "notifications") { |event| connections.each { |connection| connection.send(event) } }
loop do
action, connection = connection_channel.receive
case action
when true
connections << connection
when false
connections.delete(connection)
end
end end
end end
@ -4469,15 +4480,7 @@ get "/api/v1/auth/notifications" do |env|
topics = env.params.query["topics"]?.try &.split(",").uniq.first(1000) topics = env.params.query["topics"]?.try &.split(",").uniq.first(1000)
topics ||= [] of String topics ||= [] of String
notification_channel = Channel(PQ::Notification).new create_notification_stream(env, proxies, config, Kemal.config, decrypt_function, topics, connection_channel)
notification_channels << notification_channel
begin
create_notification_stream(env, proxies, config, Kemal.config, decrypt_function, topics, notification_channel)
rescue ex
ensure
notification_channels.delete(notification_channel)
end
end end
post "/api/v1/auth/notifications" do |env| post "/api/v1/auth/notifications" do |env|
@ -4486,15 +4489,7 @@ post "/api/v1/auth/notifications" do |env|
topics = env.params.body["topics"]?.try &.split(",").uniq.first(1000) topics = env.params.body["topics"]?.try &.split(",").uniq.first(1000)
topics ||= [] of String topics ||= [] of String
notification_channel = Channel(PQ::Notification).new create_notification_stream(env, proxies, config, Kemal.config, decrypt_function, topics, connection_channel)
notification_channels << notification_channel
begin
create_notification_stream(env, proxies, config, Kemal.config, decrypt_function, topics, notification_channel)
rescue ex
ensure
notification_channels.delete(notification_channel)
end
end end
get "/api/v1/auth/preferences" do |env| get "/api/v1/auth/preferences" do |env|

View File

@ -661,7 +661,10 @@ def copy_in_chunks(input, output, chunk_size = 4096)
end end
end end
def create_notification_stream(env, proxies, config, kemal_config, decrypt_function, topics, notification_channel) def create_notification_stream(env, proxies, config, kemal_config, decrypt_function, topics, connection_channel)
connection = Channel(PQ::Notification).new
connection_channel.send({true, connection})
locale = LOCALES[env.get("preferences").as(Preferences).locale]? locale = LOCALES[env.get("preferences").as(Preferences).locale]?
since = env.params.query["since"]?.try &.to_i? since = env.params.query["since"]?.try &.to_i?
@ -669,15 +672,87 @@ def create_notification_stream(env, proxies, config, kemal_config, decrypt_funct
if topics.includes? "debug" if topics.includes? "debug"
spawn do spawn do
begin
loop do
time_span = [0, 0, 0, 0]
time_span[rand(4)] = rand(30) + 5
published = Time.now - Time::Span.new(time_span[0], time_span[1], time_span[2], time_span[3])
video_id = TEST_IDS[rand(TEST_IDS.size)]
video = get_video(video_id, PG_DB, proxies)
video.published = published
response = JSON.parse(video.to_json(locale, config, kemal_config, decrypt_function))
if fields_text = env.params.query["fields"]?
begin
JSONFilter.filter(response, fields_text)
rescue ex
env.response.status_code = 400
response = {"error" => ex.message}
end
end
env.response.puts "id: #{id}"
env.response.puts "data: #{response.to_json}"
env.response.puts
env.response.flush
id += 1
sleep 1.minute
end
rescue ex
end
end
end
spawn do
begin
if since
topics.try &.each do |topic|
case topic
when .match(/UC[A-Za-z0-9_-]{22}/)
PG_DB.query_all("SELECT * FROM channel_videos WHERE ucid = $1 AND published > $2 ORDER BY published DESC LIMIT 15",
topic, Time.unix(since.not_nil!), as: ChannelVideo).each do |video|
response = JSON.parse(video.to_json(locale, config, Kemal.config))
if fields_text = env.params.query["fields"]?
begin
JSONFilter.filter(response, fields_text)
rescue ex
env.response.status_code = 400
response = {"error" => ex.message}
end
end
env.response.puts "id: #{id}"
env.response.puts "data: #{response.to_json}"
env.response.puts
env.response.flush
id += 1
end
else
# TODO
end
end
end
end
end
spawn do
begin
loop do loop do
time_span = [0, 0, 0, 0] event = connection.receive
time_span[rand(4)] = rand(30) + 5
published = Time.now - Time::Span.new(time_span[0], time_span[1], time_span[2], time_span[3]) notification = JSON.parse(event.payload)
video_id = TEST_IDS[rand(TEST_IDS.size)] topic = notification["topic"].as_s
video_id = notification["videoId"].as_s
published = notification["published"].as_i64
video = get_video(video_id, PG_DB, proxies) video = get_video(video_id, PG_DB, proxies)
video.published = published video.published = Time.unix(published)
response = JSON.parse(video.to_json(locale, config, kemal_config, decrypt_function)) response = JSON.parse(video.to_json(locale, config, Kemal.config, decrypt_function))
if fields_text = env.params.query["fields"]? if fields_text = env.params.query["fields"]?
begin begin
@ -688,88 +763,31 @@ def create_notification_stream(env, proxies, config, kemal_config, decrypt_funct
end end
end end
env.response.puts "id: #{id}" if topics.try &.includes? topic
env.response.puts "data: #{response.to_json}" env.response.puts "id: #{id}"
env.response.puts env.response.puts "data: #{response.to_json}"
env.response.flush env.response.puts
env.response.flush
id += 1 id += 1
sleep 1.minute
end
end
end
spawn do
if since
topics.try &.each do |topic|
case topic
when .match(/UC[A-Za-z0-9_-]{22}/)
PG_DB.query_all("SELECT * FROM channel_videos WHERE ucid = $1 AND published > $2 ORDER BY published DESC LIMIT 15",
topic, Time.unix(since.not_nil!), as: ChannelVideo).each do |video|
response = JSON.parse(video.to_json(locale, config, Kemal.config))
if fields_text = env.params.query["fields"]?
begin
JSONFilter.filter(response, fields_text)
rescue ex
env.response.status_code = 400
response = {"error" => ex.message}
end
end
env.response.puts "id: #{id}"
env.response.puts "data: #{response.to_json}"
env.response.puts
env.response.flush
id += 1
end
else
# TODO
end end
end end
rescue ex
ensure
connection_channel.send({false, connection})
end end
end end
spawn do begin
# Send heartbeat
loop do loop do
event = notification_channel.receive env.response.puts ":keepalive #{Time.now.to_unix}"
env.response.puts
notification = JSON.parse(event.payload) env.response.flush
topic = notification["topic"].as_s sleep (20 + rand(11)).seconds
video_id = notification["videoId"].as_s
published = notification["published"].as_i64
video = get_video(video_id, PG_DB, proxies)
video.published = Time.unix(published)
response = JSON.parse(video.to_json(locale, config, Kemal.config, decrypt_function))
if fields_text = env.params.query["fields"]?
begin
JSONFilter.filter(response, fields_text)
rescue ex
env.response.status_code = 400
response = {"error" => ex.message}
end
end
if topics.try &.includes? topic
env.response.puts "id: #{id}"
env.response.puts "data: #{response.to_json}"
env.response.puts
env.response.flush
id += 1
end
end end
end rescue ex
ensure
# Send heartbeat connection_channel.send({false, connection})
loop do
env.response.puts ":keepalive #{Time.now.to_unix}"
env.response.puts
env.response.flush
sleep (20 + rand(11)).seconds
end end
end end