Use frosty instead of msgpack, compress everything

This commit is contained in:
Zed 2020-06-07 08:50:40 +02:00
parent a8553db66e
commit 908da72ba9
2 changed files with 23 additions and 24 deletions

View File

@ -19,9 +19,9 @@ requires "sass"
requires "markdown#head"
requires "https://github.com/zedeus/redis#head"
requires "redpool#head"
requires "msgpack4nim >= 0.3.1"
requires "packedjson"
requires "snappy#head"
requires "https://github.com/disruptek/frosty#0.0.6"
# Tasks

View File

@ -1,6 +1,5 @@
import asyncdispatch, times, strutils, tables
import redis, redpool, msgpack4nim
export redpool, msgpack4nim
import redis, redpool, frosty, snappy
import types, api
@ -16,20 +15,26 @@ proc setCacheTimes*(cfg: Config) =
rssCacheTime = cfg.rssCacheTime * 60
listCacheTime = cfg.listCacheTime * 60
proc migrate*(key, match: string) {.async.} =
pool.withAcquire(r):
let hasKey = await r.get(key)
if hasKey == redisNil:
let list = await r.scan(newCursor(0), match, 100000)
r.startPipelining()
for item in list:
if item != "p:":
discard await r.del(item)
await r.setk(key, "true")
discard await r.flushPipeline()
proc initRedisPool*(cfg: Config) {.async.} =
try:
pool = await newRedisPool(cfg.redisConns, maxConns=cfg.redisMaxConns,
host=cfg.redisHost, port=cfg.redisPort)
pool.withAcquire(r):
let snappyRss = await r.get("snappyRss")
if snappyRss == redisNil:
let list = await r.scan(newCursor(0), "rss:*", 10000)
r.startPipelining()
for rss in list:
discard await r.del(rss)
discard await r.flushPipeline()
await r.setk("snappyRss", "true")
await migrate("snappyRss", "rss:*")
await migrate("frosty", "*")
except OSError:
echo "Failed to connect to Redis."
quit(1)
@ -37,12 +42,6 @@ proc initRedisPool*(cfg: Config) {.async.} =
template toKey(p: Profile): string = "p:" & toLower(p.username)
template toKey(l: List): string = toLower("l:" & l.username & '/' & l.name)
template to(s: string; typ: typedesc): untyped =
var res: typ
if s.len > 0:
s.unpack(res)
res
proc get(query: string): Future[string] {.async.} =
pool.withAcquire(r):
result = await r.get(query)
@ -52,16 +51,16 @@ proc setex(key: string; time: int; data: string) {.async.} =
discard await r.setex(key, time, data)
proc cache*(data: List) {.async.} =
await setex(data.toKey, listCacheTime, data.pack)
await setex(data.toKey, listCacheTime, compress(freeze(data)))
proc cache*(data: PhotoRail; id: string) {.async.} =
await setex("pr:" & id, baseCacheTime, data.pack)
await setex("pr:" & id, baseCacheTime, compress(freeze(data)))
proc cache*(data: Profile) {.async.} =
if data.username.len == 0: return
pool.withAcquire(r):
r.startPipelining()
discard await r.setex(data.toKey, baseCacheTime, pack(data))
discard await r.setex(data.toKey, baseCacheTime, compress(freeze(data)))
discard await r.hset("p:", toLower(data.username), data.id)
discard await r.flushPipeline()
@ -88,7 +87,7 @@ proc getProfileId*(username: string): Future[string] {.async.} =
proc getCachedProfile*(username: string; fetch=true): Future[Profile] {.async.} =
let prof = await get("p:" & toLower(username))
if prof != redisNil:
result = prof.to(Profile)
uncompress(prof).thaw(result)
elif fetch:
result = await getProfile(username)
@ -96,7 +95,7 @@ proc getCachedPhotoRail*(id: string): Future[PhotoRail] {.async.} =
if id.len == 0: return
let rail = await get("pr:" & toLower(id))
if rail != redisNil:
result = rail.to(PhotoRail)
uncompress(rail).thaw(result)
else:
result = await getPhotoRail(id)
await cache(result, id)
@ -106,7 +105,7 @@ proc getCachedList*(username=""; name=""; id=""): Future[List] {.async.} =
else: await get(toLower("l:" & username & '/' & name))
if list != redisNil:
result = list.to(List)
uncompress(list).thaw(result)
else:
if id.len > 0:
result = await getGraphListById(id)