Refactor connect_listen for notifications

This commit is contained in:
Omar Roth 2019-06-02 07:41:53 -05:00
parent 576067c1e5
commit 71bf8b6b4d
No known key found for this signature in database
GPG key ID: B8254FB7EC3D37F2
2 changed files with 115 additions and 90 deletions

View file

@ -186,6 +186,13 @@ spawn do
end end
end end
notification_channels = [] of Channel(PQ::Notification)
PG.connect_listen(PG_URL, "notifications") do |event|
notification_channels.each do |channel|
channel.send(event)
end
end
proxies = PROXY_LIST proxies = PROXY_LIST
before_all do |env| before_all do |env|
@ -4457,17 +4464,37 @@ get "/api/v1/mixes/:rdid" do |env|
end end
get "/api/v1/auth/notifications" do |env| get "/api/v1/auth/notifications" do |env|
env.response.content_type = "text/event-stream"
topics = env.params.query["topics"]?.try &.split(",").uniq.first(1000) topics = env.params.query["topics"]?.try &.split(",").uniq.first(1000)
topics ||= [] of String topics ||= [] of String
create_notification_stream(env, proxies, config, Kemal.config, decrypt_function, topics) notification_channel = Channel(PQ::Notification).new
notification_channels << notification_channel
begin
create_notification_stream(env, proxies, config, Kemal.config, decrypt_function, topics, notification_channel)
rescue ex
ensure
notification_channels.delete(notification_channel)
end
end end
post "/api/v1/auth/notifications" do |env| post "/api/v1/auth/notifications" do |env|
env.response.content_type = "text/event-stream"
topics = env.params.body["topics"]?.try &.split(",").uniq.first(1000) topics = env.params.body["topics"]?.try &.split(",").uniq.first(1000)
topics ||= [] of String topics ||= [] of String
create_notification_stream(env, proxies, config, Kemal.config, decrypt_function, topics) notification_channel = Channel(PQ::Notification).new
notification_channels << notification_channel
begin
create_notification_stream(env, proxies, config, Kemal.config, decrypt_function, topics, notification_channel)
rescue ex
ensure
notification_channels.delete(notification_channel)
end
end end
get "/api/v1/auth/preferences" do |env| get "/api/v1/auth/preferences" do |env|

View file

@ -661,89 +661,23 @@ def copy_in_chunks(input, output, chunk_size = 4096)
end end
end end
def create_notification_stream(env, proxies, config, kemal_config, decrypt_function, topics) def create_notification_stream(env, proxies, config, kemal_config, decrypt_function, topics, notification_channel)
locale = LOCALES[env.get("preferences").as(Preferences).locale]? locale = LOCALES[env.get("preferences").as(Preferences).locale]?
env.response.content_type = "text/event-stream"
since = env.params.query["since"]?.try &.to_i? since = env.params.query["since"]?.try &.to_i?
id = 0
begin if topics.includes? "debug"
id = 0
if topics.includes? "debug"
spawn do
loop do
time_span = [0, 0, 0, 0]
time_span[rand(4)] = rand(30) + 5
published = Time.now - Time::Span.new(time_span[0], time_span[1], time_span[2], time_span[3])
video_id = TEST_IDS[rand(TEST_IDS.size)]
video = get_video(video_id, PG_DB, proxies)
video.published = published
response = JSON.parse(video.to_json(locale, config, kemal_config, decrypt_function))
if fields_text = env.params.query["fields"]?
begin
JSONFilter.filter(response, fields_text)
rescue ex
env.response.status_code = 400
response = {"error" => ex.message}
end
end
env.response.puts "id: #{id}"
env.response.puts "data: #{response.to_json}"
env.response.puts
env.response.flush
id += 1
sleep 1.minute
end
end
end
spawn do spawn do
if since loop do
topics.try &.each do |topic| time_span = [0, 0, 0, 0]
case topic time_span[rand(4)] = rand(30) + 5
when .match(/UC[A-Za-z0-9_-]{22}/) published = Time.now - Time::Span.new(time_span[0], time_span[1], time_span[2], time_span[3])
PG_DB.query_all("SELECT * FROM channel_videos WHERE ucid = $1 AND published > $2 ORDER BY published DESC LIMIT 15", video_id = TEST_IDS[rand(TEST_IDS.size)]
topic, Time.unix(since.not_nil!), as: ChannelVideo).each do |video|
response = JSON.parse(video.to_json(locale, config, Kemal.config))
if fields_text = env.params.query["fields"]?
begin
JSONFilter.filter(response, fields_text)
rescue ex
env.response.status_code = 400
response = {"error" => ex.message}
end
end
env.response.puts "id: #{id}"
env.response.puts "data: #{response.to_json}"
env.response.puts
env.response.flush
id += 1
end
else
# TODO
end
end
end
PG.connect_listen(PG_URL, "notifications") do |event|
notification = JSON.parse(event.payload)
topic = notification["topic"].as_s
video_id = notification["videoId"].as_s
published = notification["published"].as_i64
video = get_video(video_id, PG_DB, proxies) video = get_video(video_id, PG_DB, proxies)
video.published = Time.unix(published) video.published = published
response = JSON.parse(video.to_json(locale, config, Kemal.config, decrypt_function)) response = JSON.parse(video.to_json(locale, config, kemal_config, decrypt_function))
if fields_text = env.params.query["fields"]? if fields_text = env.params.query["fields"]?
begin begin
@ -754,24 +688,88 @@ def create_notification_stream(env, proxies, config, kemal_config, decrypt_funct
end end
end end
if topics.try &.includes? topic env.response.puts "id: #{id}"
env.response.puts "id: #{id}" env.response.puts "data: #{response.to_json}"
env.response.puts "data: #{response.to_json}" env.response.puts
env.response.puts env.response.flush
env.response.flush
id += 1 id += 1
sleep 1.minute
end
end
end
spawn do
if since
topics.try &.each do |topic|
case topic
when .match(/UC[A-Za-z0-9_-]{22}/)
PG_DB.query_all("SELECT * FROM channel_videos WHERE ucid = $1 AND published > $2 ORDER BY published DESC LIMIT 15",
topic, Time.unix(since.not_nil!), as: ChannelVideo).each do |video|
response = JSON.parse(video.to_json(locale, config, Kemal.config))
if fields_text = env.params.query["fields"]?
begin
JSONFilter.filter(response, fields_text)
rescue ex
env.response.status_code = 400
response = {"error" => ex.message}
end
end
env.response.puts "id: #{id}"
env.response.puts "data: #{response.to_json}"
env.response.puts
env.response.flush
id += 1
end
else
# TODO
end end
end end
end end
end
# Send heartbeat spawn do
loop do loop do
env.response.puts ":keepalive #{Time.now.to_unix}" event = notification_channel.receive
env.response.puts
env.response.flush notification = JSON.parse(event.payload)
sleep (20 + rand(11)).seconds topic = notification["topic"].as_s
video_id = notification["videoId"].as_s
published = notification["published"].as_i64
video = get_video(video_id, PG_DB, proxies)
video.published = Time.unix(published)
response = JSON.parse(video.to_json(locale, config, Kemal.config, decrypt_function))
if fields_text = env.params.query["fields"]?
begin
JSONFilter.filter(response, fields_text)
rescue ex
env.response.status_code = 400
response = {"error" => ex.message}
end
end
if topics.try &.includes? topic
env.response.puts "id: #{id}"
env.response.puts "data: #{response.to_json}"
env.response.puts
env.response.flush
id += 1
end
end end
rescue end
# Send heartbeat
loop do
env.response.puts ":keepalive #{Time.now.to_unix}"
env.response.puts
env.response.flush
sleep (20 + rand(11)).seconds
end end
end end