2017-10-11 04:02:16 +02:00
|
|
|
package ytsync
|
|
|
|
|
|
|
|
import (
|
|
|
|
"bytes"
|
|
|
|
"encoding/json"
|
|
|
|
"io/ioutil"
|
|
|
|
"net/http"
|
|
|
|
"os"
|
|
|
|
"sort"
|
|
|
|
"strconv"
|
|
|
|
"strings"
|
|
|
|
"sync"
|
|
|
|
"sync/atomic"
|
|
|
|
"time"
|
|
|
|
|
|
|
|
"github.com/lbryio/lbry.go/jsonrpc"
|
|
|
|
|
|
|
|
"github.com/garyburd/redigo/redis"
|
|
|
|
"github.com/go-errors/errors"
|
|
|
|
ytdl "github.com/kkdai/youtube"
|
2017-11-02 16:20:22 +01:00
|
|
|
"github.com/shopspring/decimal"
|
2017-10-11 04:02:16 +02:00
|
|
|
log "github.com/sirupsen/logrus"
|
|
|
|
"google.golang.org/api/googleapi/transport"
|
|
|
|
"google.golang.org/api/youtube/v3"
|
|
|
|
)
|
|
|
|
|
|
|
|
const (
|
2017-10-18 02:35:19 +02:00
|
|
|
redisHashKey = "ytsync"
|
|
|
|
redisSyncedVal = "t"
|
|
|
|
channelClaimAmount = 0.01
|
|
|
|
publishAmount = 0.01
|
2017-10-11 04:02:16 +02:00
|
|
|
)
|
|
|
|
|
|
|
|
// Sync stores the options that control how syncing happens
|
|
|
|
type Sync struct {
|
|
|
|
YoutubeAPIKey string
|
|
|
|
YoutubeChannelID string
|
|
|
|
LbryChannelName string
|
|
|
|
StopOnError bool
|
|
|
|
MaxTries int
|
|
|
|
ConcurrentVideos int
|
|
|
|
|
|
|
|
daemon *jsonrpc.Client
|
|
|
|
claimAddress string
|
|
|
|
videoDirectory string
|
|
|
|
redisPool *redis.Pool
|
|
|
|
}
|
|
|
|
|
2017-10-11 19:13:47 +02:00
|
|
|
func (s *Sync) init() error {
|
2017-10-11 04:02:16 +02:00
|
|
|
var err error
|
|
|
|
|
|
|
|
s.redisPool = &redis.Pool{
|
|
|
|
MaxIdle: 3,
|
|
|
|
IdleTimeout: 5 * time.Minute,
|
|
|
|
Dial: func() (redis.Conn, error) { return redis.Dial("tcp", ":6379") },
|
|
|
|
TestOnBorrow: func(c redis.Conn, t time.Time) error {
|
|
|
|
if time.Since(t) < time.Minute {
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
_, err := c.Do("PING")
|
|
|
|
return err
|
|
|
|
},
|
|
|
|
}
|
|
|
|
|
|
|
|
s.videoDirectory, err = ioutil.TempDir("", "ytsync")
|
|
|
|
if err != nil {
|
2017-11-02 16:20:22 +01:00
|
|
|
return errors.Wrap(err, 0)
|
2017-10-11 04:02:16 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
s.daemon = jsonrpc.NewClient("")
|
|
|
|
|
|
|
|
addresses, err := s.daemon.WalletList()
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
} else if addresses == nil || len(*addresses) == 0 {
|
|
|
|
return errors.New("could not find an address in wallet")
|
|
|
|
}
|
2017-10-11 18:45:56 +02:00
|
|
|
s.claimAddress = (*addresses)[0]
|
|
|
|
if s.claimAddress == "" {
|
2017-10-11 04:02:16 +02:00
|
|
|
return errors.New("found blank claim address")
|
|
|
|
}
|
|
|
|
|
2017-11-02 16:20:22 +01:00
|
|
|
err = s.ensureEnoughUTXOs()
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
2017-10-11 19:13:47 +02:00
|
|
|
if s.LbryChannelName != "" {
|
|
|
|
err = s.ensureChannelOwnership()
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
2017-11-02 16:20:22 +01:00
|
|
|
func (s *Sync) CountVideos() (uint64, error) {
|
|
|
|
client := &http.Client{
|
|
|
|
Transport: &transport.APIKey{Key: s.YoutubeAPIKey},
|
|
|
|
}
|
|
|
|
|
|
|
|
service, err := youtube.New(client)
|
|
|
|
if err != nil {
|
|
|
|
return 0, errors.WrapPrefix(err, "error creating YouTube service", 0)
|
|
|
|
}
|
|
|
|
|
|
|
|
response, err := service.Channels.List("statistics").Id(s.YoutubeChannelID).Do()
|
|
|
|
if err != nil {
|
|
|
|
return 0, errors.WrapPrefix(err, "error getting channels", 0)
|
|
|
|
}
|
|
|
|
|
|
|
|
if len(response.Items) < 1 {
|
|
|
|
return 0, errors.New("youtube channel not found")
|
|
|
|
}
|
|
|
|
|
|
|
|
return response.Items[0].Statistics.VideoCount, nil
|
|
|
|
}
|
|
|
|
|
2017-10-11 19:13:47 +02:00
|
|
|
func (s *Sync) Go() error {
|
|
|
|
var err error
|
|
|
|
|
|
|
|
err = s.init()
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
2017-10-11 04:02:16 +02:00
|
|
|
var wg sync.WaitGroup
|
|
|
|
videoQueue := make(chan video)
|
|
|
|
|
2017-10-11 19:13:47 +02:00
|
|
|
queueStopChan := make(chan struct{})
|
2017-10-11 04:02:16 +02:00
|
|
|
sendStopEnqueuing := sync.Once{}
|
|
|
|
|
|
|
|
var videoErrored atomic.Value
|
|
|
|
videoErrored.Store(false)
|
|
|
|
if s.StopOnError {
|
|
|
|
log.Println("Will stop publishing if an error is detected")
|
|
|
|
}
|
|
|
|
|
|
|
|
for i := 0; i < s.ConcurrentVideos; i++ {
|
|
|
|
go func() {
|
|
|
|
wg.Add(1)
|
|
|
|
defer wg.Done()
|
|
|
|
|
|
|
|
for {
|
|
|
|
v, more := <-videoQueue
|
|
|
|
if !more {
|
|
|
|
return
|
|
|
|
}
|
|
|
|
if s.StopOnError && videoErrored.Load().(bool) {
|
|
|
|
log.Println("Video errored. Exiting")
|
|
|
|
return
|
|
|
|
}
|
|
|
|
|
2017-10-11 19:13:47 +02:00
|
|
|
log.Println("========================================")
|
|
|
|
|
2017-10-11 04:02:16 +02:00
|
|
|
tryCount := 0
|
|
|
|
for {
|
|
|
|
tryCount++
|
|
|
|
err := s.processVideo(v)
|
|
|
|
|
|
|
|
if err != nil {
|
|
|
|
log.Errorln("error processing video: " + err.Error())
|
|
|
|
if s.StopOnError {
|
|
|
|
videoErrored.Store(true)
|
|
|
|
sendStopEnqueuing.Do(func() {
|
2017-10-11 19:13:47 +02:00
|
|
|
queueStopChan <- struct{}{}
|
2017-10-11 04:02:16 +02:00
|
|
|
})
|
|
|
|
} else if s.MaxTries > 1 {
|
|
|
|
if strings.Contains(err.Error(), "non 200 status code received") ||
|
|
|
|
strings.Contains(err.Error(), " reason: 'This video contains content from") {
|
|
|
|
log.Println("This error should not be retried at all")
|
|
|
|
} else if tryCount >= s.MaxTries {
|
2017-11-02 16:20:22 +01:00
|
|
|
log.Println("Video failed after " + strconv.Itoa(s.MaxTries) + " retries, exiting")
|
|
|
|
videoErrored.Store(true)
|
|
|
|
sendStopEnqueuing.Do(func() {
|
|
|
|
queueStopChan <- struct{}{}
|
|
|
|
})
|
2017-10-11 04:02:16 +02:00
|
|
|
} else {
|
|
|
|
log.Println("Retrying")
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
break
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}()
|
|
|
|
}
|
|
|
|
|
2017-10-11 19:13:47 +02:00
|
|
|
err = s.enqueueVideosFromChannel(s.YoutubeChannelID, &videoQueue, &queueStopChan)
|
2017-10-11 04:02:16 +02:00
|
|
|
close(videoQueue)
|
|
|
|
wg.Wait()
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
2017-11-02 16:20:22 +01:00
|
|
|
func (s *Sync) ensureEnoughUTXOs() error {
|
|
|
|
utxolist, err := s.daemon.UTXOList()
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
} else if utxolist == nil {
|
|
|
|
return errors.New("no response")
|
|
|
|
}
|
|
|
|
|
|
|
|
target := 50
|
|
|
|
count := 0
|
|
|
|
|
|
|
|
for _, utxo := range *utxolist {
|
|
|
|
if !utxo.IsClaim && !utxo.IsSupport && !utxo.IsUpdate && utxo.Amount.Cmp(decimal.New(0, 0)) == 1 {
|
|
|
|
count++
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if count < target {
|
|
|
|
newAddresses := target - count
|
|
|
|
|
|
|
|
balance, err := s.daemon.WalletBalance()
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
} else if balance == nil {
|
|
|
|
return errors.New("no response")
|
|
|
|
}
|
|
|
|
|
|
|
|
amountPerAddress := decimal.Decimal(*balance).Div(decimal.NewFromFloat(float64(target)))
|
|
|
|
log.Infof("Putting %s credits into each of %d new addresses", amountPerAddress.String(), newAddresses)
|
|
|
|
prefillTx, err := s.daemon.WalletPrefillAddresses(newAddresses, amountPerAddress, true)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
} else if prefillTx == nil {
|
|
|
|
return errors.New("no response")
|
|
|
|
} else if !prefillTx.Complete || !prefillTx.Broadcast {
|
|
|
|
return errors.New("failed to prefill addresses")
|
|
|
|
}
|
|
|
|
|
|
|
|
wait := 15 * time.Second
|
|
|
|
log.Println("Waiting " + wait.String() + " for lbryum to let us know we have the new addresses")
|
|
|
|
time.Sleep(wait)
|
|
|
|
|
|
|
|
log.Println("Creating UTXOs and waiting for them to be confirmed")
|
|
|
|
s.waitUntilUTXOsConfirmed()
|
|
|
|
}
|
|
|
|
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
func (s *Sync) waitUntilUTXOsConfirmed() error {
|
|
|
|
for {
|
|
|
|
r, err := s.daemon.UTXOList()
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
} else if r == nil {
|
|
|
|
return errors.New("no response")
|
|
|
|
}
|
|
|
|
|
|
|
|
allConfirmed := true
|
|
|
|
for _, utxo := range *r {
|
|
|
|
if utxo.Height == 0 {
|
|
|
|
allConfirmed = false
|
|
|
|
break
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if allConfirmed {
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
wait := 30 * time.Second
|
|
|
|
log.Println("Waiting " + wait.String() + "...")
|
|
|
|
time.Sleep(wait)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2017-10-11 04:02:16 +02:00
|
|
|
func (s *Sync) ensureChannelOwnership() error {
|
|
|
|
channels, err := s.daemon.ChannelListMine()
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
} else if channels == nil {
|
|
|
|
return errors.New("no channels")
|
|
|
|
}
|
|
|
|
|
|
|
|
for _, channel := range *channels {
|
|
|
|
if channel.Name == s.LbryChannelName {
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
resolveResp, err := s.daemon.Resolve(s.LbryChannelName)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
2017-10-11 17:25:43 +02:00
|
|
|
channelNotFound := (*resolveResp)[s.LbryChannelName].Error != nil && strings.Contains(*((*resolveResp)[s.LbryChannelName].Error), "cannot be resolved")
|
2017-10-11 04:02:16 +02:00
|
|
|
|
|
|
|
if !channelNotFound {
|
|
|
|
return errors.New("Channel exists and we don't own it. Pick another channel.")
|
|
|
|
}
|
|
|
|
|
2017-10-18 02:35:19 +02:00
|
|
|
_, err = s.daemon.ChannelNew(s.LbryChannelName, channelClaimAmount)
|
2017-10-11 04:02:16 +02:00
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
// niko's code says "unfortunately the queues in the daemon are not yet merged so we must give it some time for the channel to go through"
|
|
|
|
wait := 15 * time.Second
|
|
|
|
log.Println("Waiting " + wait.String() + " for channel claim to go through")
|
|
|
|
time.Sleep(wait)
|
|
|
|
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
2017-10-11 19:13:47 +02:00
|
|
|
func (s *Sync) enqueueVideosFromChannel(channelID string, videoChan *chan video, queueStopChan *chan struct{}) error {
|
2017-10-11 04:02:16 +02:00
|
|
|
client := &http.Client{
|
|
|
|
Transport: &transport.APIKey{Key: s.YoutubeAPIKey},
|
|
|
|
}
|
|
|
|
|
|
|
|
service, err := youtube.New(client)
|
|
|
|
if err != nil {
|
|
|
|
return errors.WrapPrefix(err, "error creating YouTube service", 0)
|
|
|
|
}
|
|
|
|
|
|
|
|
response, err := service.Channels.List("contentDetails").Id(channelID).Do()
|
|
|
|
if err != nil {
|
|
|
|
return errors.WrapPrefix(err, "error getting channels", 0)
|
|
|
|
}
|
|
|
|
|
|
|
|
if len(response.Items) < 1 {
|
|
|
|
return errors.New("youtube channel not found")
|
|
|
|
}
|
|
|
|
|
|
|
|
if response.Items[0].ContentDetails.RelatedPlaylists == nil {
|
|
|
|
return errors.New("no related playlists")
|
|
|
|
}
|
|
|
|
|
|
|
|
playlistID := response.Items[0].ContentDetails.RelatedPlaylists.Uploads
|
|
|
|
if playlistID == "" {
|
|
|
|
return errors.New("no channel playlist")
|
|
|
|
}
|
|
|
|
|
|
|
|
videos := []video{}
|
|
|
|
|
|
|
|
nextPageToken := ""
|
|
|
|
for {
|
|
|
|
req := service.PlaylistItems.List("snippet").
|
|
|
|
PlaylistId(playlistID).
|
|
|
|
MaxResults(50).
|
|
|
|
PageToken(nextPageToken)
|
|
|
|
|
|
|
|
playlistResponse, err := req.Do()
|
|
|
|
if err != nil {
|
|
|
|
return errors.WrapPrefix(err, "error getting playlist items", 0)
|
|
|
|
}
|
|
|
|
|
|
|
|
if len(playlistResponse.Items) < 1 {
|
|
|
|
return errors.New("playlist items not found")
|
|
|
|
}
|
|
|
|
|
|
|
|
for _, item := range playlistResponse.Items {
|
|
|
|
// todo: there's thumbnail info here. why did we need lambda???
|
|
|
|
publishedAt, err := time.Parse(time.RFC3339Nano, item.Snippet.PublishedAt)
|
|
|
|
if err != nil {
|
|
|
|
return errors.WrapPrefix(err, "failed to parse time", 0)
|
|
|
|
}
|
|
|
|
|
|
|
|
// normally we'd send the video into the channel here, but youtube api doesn't have sorting
|
|
|
|
// so we have to get ALL the videos, then sort them, then send them in
|
|
|
|
videos = append(videos, video{
|
|
|
|
id: item.Snippet.ResourceId.VideoId,
|
|
|
|
channelID: channelID,
|
|
|
|
title: item.Snippet.Title,
|
|
|
|
description: item.Snippet.Description,
|
|
|
|
channelTitle: item.Snippet.ChannelTitle,
|
|
|
|
playlistPosition: item.Snippet.Position,
|
|
|
|
publishedAt: publishedAt,
|
|
|
|
dir: s.videoDirectory,
|
|
|
|
})
|
|
|
|
}
|
|
|
|
|
|
|
|
log.Infoln("Got info for " + strconv.Itoa(len(videos)) + " videos from youtube API")
|
|
|
|
|
|
|
|
nextPageToken = playlistResponse.NextPageToken
|
|
|
|
if nextPageToken == "" {
|
|
|
|
break
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
sort.Sort(byPublishedAt(videos))
|
|
|
|
//or sort.Sort(sort.Reverse(byPlaylistPosition(videos)))
|
|
|
|
|
2017-10-11 19:13:47 +02:00
|
|
|
Enqueue:
|
2017-10-11 04:02:16 +02:00
|
|
|
for _, v := range videos {
|
|
|
|
select {
|
|
|
|
case *videoChan <- v:
|
2017-10-11 19:13:47 +02:00
|
|
|
case <-*queueStopChan:
|
|
|
|
break Enqueue
|
2017-10-11 04:02:16 +02:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
func (s *Sync) processVideo(v video) error {
|
|
|
|
log.Println("Processing " + v.id + " (" + strconv.Itoa(int(v.playlistPosition)) + " in channel)")
|
2017-11-02 16:20:22 +01:00
|
|
|
defer func(start time.Time) {
|
|
|
|
log.Println(v.id + " took " + time.Since(start).String())
|
|
|
|
}(time.Now())
|
2017-10-11 04:02:16 +02:00
|
|
|
|
|
|
|
conn := s.redisPool.Get()
|
|
|
|
defer conn.Close()
|
|
|
|
|
|
|
|
alreadyPublished, err := redis.String(conn.Do("HGET", redisHashKey, v.id))
|
|
|
|
if err != nil && err != redis.ErrNil {
|
|
|
|
return errors.WrapPrefix(err, "redis error", 0)
|
|
|
|
|
|
|
|
}
|
|
|
|
if alreadyPublished == redisSyncedVal {
|
|
|
|
log.Println(v.id + " already published")
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
//download and thumbnail can be done in parallel
|
|
|
|
err = downloadVideo(v)
|
|
|
|
if err != nil {
|
|
|
|
return errors.WrapPrefix(err, "download error", 0)
|
|
|
|
}
|
|
|
|
|
|
|
|
err = triggerThumbnailSave(v.id)
|
|
|
|
if err != nil {
|
|
|
|
return errors.WrapPrefix(err, "thumbnail error", 0)
|
|
|
|
}
|
|
|
|
|
|
|
|
err = s.publish(v, conn)
|
|
|
|
if err != nil {
|
|
|
|
return errors.WrapPrefix(err, "publish error", 0)
|
|
|
|
}
|
|
|
|
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
func downloadVideo(v video) error {
|
|
|
|
verbose := false
|
|
|
|
videoPath := v.getFilename()
|
|
|
|
|
|
|
|
_, err := os.Stat(videoPath)
|
|
|
|
if err != nil && !os.IsNotExist(err) {
|
|
|
|
return err
|
|
|
|
} else if err == nil {
|
|
|
|
log.Println(v.id + " already exists at " + videoPath)
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
downloader := ytdl.NewYoutube(verbose)
|
|
|
|
err = downloader.DecodeURL("https://www.youtube.com/watch?v=" + v.id)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
err = downloader.StartDownload(videoPath)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
log.Debugln("Downloaded " + v.id)
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
func triggerThumbnailSave(videoID string) error {
|
|
|
|
client := &http.Client{Timeout: 30 * time.Second}
|
|
|
|
|
|
|
|
params, err := json.Marshal(map[string]string{"videoid": videoID})
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
request, err := http.NewRequest(http.MethodPut, "https://jgp4g1qoud.execute-api.us-east-1.amazonaws.com/prod/thumbnail", bytes.NewBuffer(params))
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
response, err := client.Do(request)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
defer response.Body.Close()
|
|
|
|
|
|
|
|
contents, err := ioutil.ReadAll(response.Body)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
var decoded struct {
|
|
|
|
error int `json:"error"`
|
|
|
|
url string `json:"url,omitempty"`
|
|
|
|
message string `json:"message,omitempty"`
|
|
|
|
}
|
|
|
|
err = json.Unmarshal(contents, &decoded)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
if decoded.error != 0 {
|
|
|
|
return errors.New("error creating thumbnail: " + decoded.message)
|
|
|
|
}
|
|
|
|
|
|
|
|
log.Debugln("Created thumbnail for " + videoID)
|
|
|
|
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
func strPtr(s string) *string { return &s }
|
|
|
|
|
|
|
|
func (s *Sync) publish(v video, conn redis.Conn) error {
|
|
|
|
options := jsonrpc.PublishOptions{
|
|
|
|
Title: &v.title,
|
|
|
|
Author: &v.channelTitle,
|
|
|
|
Description: strPtr(v.getAbbrevDescription() + "\nhttps://www.youtube.com/watch?v=" + v.id),
|
|
|
|
Language: strPtr("en"),
|
|
|
|
ClaimAddress: &s.claimAddress,
|
|
|
|
Thumbnail: strPtr("http://berk.ninja/thumbnails/" + v.id),
|
|
|
|
License: strPtr("Copyrighted (contact author)"),
|
|
|
|
}
|
|
|
|
if s.LbryChannelName != "" {
|
|
|
|
options.ChannelName = &s.LbryChannelName
|
|
|
|
}
|
|
|
|
|
2017-10-18 02:35:19 +02:00
|
|
|
_, err := s.daemon.Publish(v.getClaimName(), v.getFilename(), publishAmount, options)
|
2017-10-11 04:02:16 +02:00
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
_, err = redis.Bool(conn.Do("HSET", redisHashKey, v.id, redisSyncedVal))
|
|
|
|
if err != nil {
|
|
|
|
return errors.New("redis error: " + err.Error())
|
|
|
|
}
|
|
|
|
|
|
|
|
return nil
|
|
|
|
}
|