sink_node.go 13 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464
  1. // Copyright 2021 EMQ Technologies Co., Ltd.
  2. //
  3. // Licensed under the Apache License, Version 2.0 (the "License");
  4. // you may not use this file except in compliance with the License.
  5. // You may obtain a copy of the License at
  6. //
  7. // http://www.apache.org/licenses/LICENSE-2.0
  8. //
  9. // Unless required by applicable law or agreed to in writing, software
  10. // distributed under the License is distributed on an "AS IS" BASIS,
  11. // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
  12. // See the License for the specific language governing permissions and
  13. // limitations under the License.
  14. package node
  15. import (
  16. "bytes"
  17. "encoding/json"
  18. "fmt"
  19. "github.com/lf-edge/ekuiper/internal/binder/io"
  20. "github.com/lf-edge/ekuiper/internal/conf"
  21. ct "github.com/lf-edge/ekuiper/internal/template"
  22. "github.com/lf-edge/ekuiper/pkg/api"
  23. "github.com/lf-edge/ekuiper/pkg/cast"
  24. "sync"
  25. "text/template"
  26. "time"
  27. )
  28. type SinkNode struct {
  29. *defaultSinkNode
  30. //static
  31. sinkType string
  32. mutex sync.RWMutex
  33. //configs (also static for sinks)
  34. options map[string]interface{}
  35. isMock bool
  36. //states varies after restart
  37. sinks []api.Sink
  38. tch chan struct{} //channel to trigger cache saved, will be trigger by checkpoint only
  39. }
  40. func NewSinkNode(name string, sinkType string, props map[string]interface{}) *SinkNode {
  41. bufferLength := 1024
  42. if c, ok := props["bufferLength"]; ok {
  43. if t, err := cast.ToInt(c, cast.STRICT); err != nil || t <= 0 {
  44. //invalid property bufferLength
  45. } else {
  46. bufferLength = t
  47. }
  48. }
  49. return &SinkNode{
  50. defaultSinkNode: &defaultSinkNode{
  51. input: make(chan interface{}, bufferLength),
  52. defaultNode: &defaultNode{
  53. name: name,
  54. concurrency: 1,
  55. ctx: nil,
  56. },
  57. },
  58. sinkType: sinkType,
  59. options: props,
  60. }
  61. }
  62. //Only for mock source, do not use it in production
  63. func NewSinkNodeWithSink(name string, sink api.Sink, props map[string]interface{}) *SinkNode {
  64. return &SinkNode{
  65. defaultSinkNode: &defaultSinkNode{
  66. input: make(chan interface{}, 1024),
  67. defaultNode: &defaultNode{
  68. name: name,
  69. concurrency: 1,
  70. ctx: nil,
  71. },
  72. },
  73. sinks: []api.Sink{sink},
  74. options: props,
  75. isMock: true,
  76. }
  77. }
  78. func (m *SinkNode) Open(ctx api.StreamContext, result chan<- error) {
  79. m.ctx = ctx
  80. logger := ctx.GetLogger()
  81. logger.Debugf("open sink node %s", m.name)
  82. if m.qos >= api.AtLeastOnce {
  83. m.tch = make(chan struct{})
  84. }
  85. go func() {
  86. if c, ok := m.options["concurrency"]; ok {
  87. if t, err := cast.ToInt(c, cast.STRICT); err != nil || t <= 0 {
  88. logger.Warnf("invalid type for concurrency property, should be positive integer but found %t", c)
  89. } else {
  90. m.concurrency = t
  91. }
  92. }
  93. runAsync := false
  94. if c, ok := m.options["runAsync"]; ok {
  95. if t, ok := c.(bool); !ok {
  96. logger.Warnf("invalid type for runAsync property, should be bool but found %t", c)
  97. } else {
  98. runAsync = t
  99. }
  100. }
  101. retryInterval := 1000
  102. if c, ok := m.options["retryInterval"]; ok {
  103. if t, err := cast.ToInt(c, cast.STRICT); err != nil || t < 0 {
  104. logger.Warnf("invalid type for retryInterval property, should be positive integer but found %t", c)
  105. } else {
  106. retryInterval = t
  107. }
  108. }
  109. retryCount := 3
  110. if c, ok := m.options["retryCount"]; ok {
  111. if t, err := cast.ToInt(c, cast.STRICT); err != nil || t < 0 {
  112. logger.Warnf("invalid type for retryCount property, should be positive integer but found %t", c)
  113. } else {
  114. retryCount = t
  115. }
  116. }
  117. cacheLength := 1024
  118. if c, ok := m.options["cacheLength"]; ok {
  119. if t, err := cast.ToInt(c, cast.STRICT); err != nil || t < 0 {
  120. logger.Warnf("invalid type for cacheLength property, should be positive integer but found %t", c)
  121. } else {
  122. cacheLength = t
  123. }
  124. }
  125. cacheSaveInterval := 1000
  126. if c, ok := m.options["cacheSaveInterval"]; ok {
  127. if t, err := cast.ToInt(c, cast.STRICT); err != nil || t < 0 {
  128. logger.Warnf("invalid type for cacheSaveInterval property, should be positive integer but found %t", c)
  129. } else {
  130. cacheSaveInterval = t
  131. }
  132. }
  133. omitIfEmpty := false
  134. if c, ok := m.options["omitIfEmpty"]; ok {
  135. if t, ok := c.(bool); !ok {
  136. logger.Warnf("invalid type for omitIfEmpty property, should be a bool value 'true/false'.", c)
  137. } else {
  138. omitIfEmpty = t
  139. }
  140. }
  141. sendSingle := false
  142. if c, ok := m.options["sendSingle"]; ok {
  143. if t, ok := c.(bool); !ok {
  144. logger.Warnf("invalid type for sendSingle property, should be a bool value 'true/false'.", c)
  145. } else {
  146. sendSingle = t
  147. }
  148. }
  149. var tp *template.Template = nil
  150. if c, ok := m.options["dataTemplate"]; ok {
  151. if t, ok := c.(string); !ok {
  152. logger.Warnf("invalid type for dateTemplate property, should be a string value.", c)
  153. } else {
  154. temp, err := template.New("sink").Funcs(ct.FuncMap).Parse(t)
  155. if err != nil {
  156. msg := fmt.Sprintf("property dataTemplate %v is invalid: %v", t, err)
  157. logger.Warnf(msg)
  158. result <- fmt.Errorf(msg)
  159. return
  160. } else {
  161. tp = temp
  162. }
  163. }
  164. }
  165. m.reset()
  166. logger.Infof("open sink node %d instances", m.concurrency)
  167. for i := 0; i < m.concurrency; i++ { // workers
  168. go func(instance int) {
  169. var sink api.Sink
  170. var err error
  171. if !m.isMock {
  172. logger.Debugf("Trying to get sink for rule %s with options %v\n", ctx.GetRuleId(), m.options)
  173. sink, err = getSink(m.sinkType, m.options)
  174. if err != nil {
  175. m.drainError(result, err, ctx, logger)
  176. return
  177. }
  178. logger.Debugf("Successfully get the sink %s", m.sinkType)
  179. m.mutex.Lock()
  180. m.sinks = append(m.sinks, sink)
  181. m.mutex.Unlock()
  182. logger.Debugf("Now is to open sink for rule %s.\n", ctx.GetRuleId())
  183. if err := sink.Open(ctx); err != nil {
  184. m.drainError(result, err, ctx, logger)
  185. return
  186. }
  187. logger.Debugf("Successfully open sink for rule %s.\n", ctx.GetRuleId())
  188. } else {
  189. sink = m.sinks[instance]
  190. }
  191. stats, err := NewStatManager("sink", ctx)
  192. if err != nil {
  193. m.drainError(result, err, ctx, logger)
  194. return
  195. }
  196. m.mutex.Lock()
  197. m.statManagers = append(m.statManagers, stats)
  198. m.mutex.Unlock()
  199. if conf.Config.Sink.DisableCache {
  200. for {
  201. select {
  202. case data := <-m.input:
  203. if newdata, processed := m.preprocess(data); processed {
  204. break
  205. } else {
  206. data = newdata
  207. }
  208. stats.SetBufferLength(int64(len(m.input)))
  209. if runAsync {
  210. go doCollect(sink, data, stats, omitIfEmpty, sendSingle, tp, ctx)
  211. } else {
  212. doCollect(sink, data, stats, omitIfEmpty, sendSingle, tp, ctx)
  213. }
  214. case <-ctx.Done():
  215. logger.Infof("sink node %s instance %d done", m.name, instance)
  216. if err := sink.Close(ctx); err != nil {
  217. logger.Warnf("close sink node %s instance %d fails: %v", m.name, instance, err)
  218. }
  219. return
  220. case <-m.tch:
  221. logger.Debugf("rule %s sink receive checkpoint, do nothing", ctx.GetRuleId())
  222. }
  223. }
  224. } else {
  225. logger.Infof("Creating sink cache")
  226. var cache *Cache
  227. if m.qos >= api.AtLeastOnce {
  228. cache = NewCheckpointbasedCache(m.input, cacheLength, m.tch, result, ctx)
  229. } else {
  230. cache = NewTimebasedCache(m.input, cacheLength, cacheSaveInterval, result, ctx)
  231. }
  232. for {
  233. select {
  234. case data := <-cache.Out:
  235. if newdata, processed := m.preprocess(data.data); processed {
  236. break
  237. } else {
  238. data.data = newdata
  239. }
  240. stats.SetBufferLength(int64(len(m.input)))
  241. if runAsync {
  242. go doCollectCacheTuple(sink, data, stats, retryInterval, retryCount, omitIfEmpty, sendSingle, tp, cache.Complete, ctx)
  243. } else {
  244. doCollectCacheTuple(sink, data, stats, retryInterval, retryCount, omitIfEmpty, sendSingle, tp, cache.Complete, ctx)
  245. }
  246. case <-ctx.Done():
  247. logger.Infof("sink node %s instance %d done", m.name, instance)
  248. if err := sink.Close(ctx); err != nil {
  249. logger.Warnf("close sink node %s instance %d fails: %v", m.name, instance, err)
  250. }
  251. return
  252. }
  253. }
  254. }
  255. }(i)
  256. }
  257. }()
  258. }
  259. func (m *SinkNode) reset() {
  260. if !m.isMock {
  261. m.sinks = nil
  262. }
  263. m.statManagers = nil
  264. }
  265. func extractInput(v []byte) ([]map[string]interface{}, error) {
  266. var j []map[string]interface{}
  267. if err := json.Unmarshal(v, &j); err != nil {
  268. return nil, fmt.Errorf("fail to decode the input %s as json: %v", v, err)
  269. }
  270. return j, nil
  271. }
  272. func doCollect(sink api.Sink, item interface{}, stats StatManager, omitIfEmpty bool, sendSingle bool, tp *template.Template, ctx api.StreamContext) {
  273. stats.IncTotalRecordsIn()
  274. stats.ProcessTimeStart()
  275. defer stats.ProcessTimeEnd()
  276. logger := ctx.GetLogger()
  277. outdatas := getOutData(stats, ctx, item, omitIfEmpty, sendSingle, tp)
  278. for _, outdata := range outdatas {
  279. if err := sink.Collect(ctx, outdata); err != nil {
  280. stats.IncTotalExceptions()
  281. logger.Warnf("sink node %s instance %d publish %s error: %v", ctx.GetOpId(), ctx.GetInstanceId(), outdata, err)
  282. } else {
  283. stats.IncTotalRecordsOut()
  284. }
  285. }
  286. }
  287. func getOutData(stats StatManager, ctx api.StreamContext, item interface{}, omitIfEmpty bool, sendSingle bool, tp *template.Template) [][]byte {
  288. logger := ctx.GetLogger()
  289. var outdatas [][]byte
  290. switch val := item.(type) {
  291. case []byte:
  292. if omitIfEmpty && string(val) == "[{}]" {
  293. return nil
  294. }
  295. var (
  296. err error
  297. j []map[string]interface{}
  298. )
  299. if sendSingle || tp != nil {
  300. j, err = extractInput(val)
  301. if err != nil {
  302. logger.Warnf("sink node %s instance %d publish %s error: %v", ctx.GetOpId(), ctx.GetInstanceId(), val, err)
  303. stats.IncTotalExceptions()
  304. return nil
  305. }
  306. logger.Debugf("receive %d records", len(j))
  307. }
  308. if !sendSingle {
  309. if tp != nil {
  310. var output bytes.Buffer
  311. err := tp.Execute(&output, j)
  312. if err != nil {
  313. logger.Warnf("sink node %s instance %d publish %s decode template error: %v", ctx.GetOpId(), ctx.GetInstanceId(), val, err)
  314. stats.IncTotalExceptions()
  315. return nil
  316. }
  317. outdatas = append(outdatas, output.Bytes())
  318. } else {
  319. outdatas = [][]byte{val}
  320. }
  321. } else {
  322. for _, r := range j {
  323. if tp != nil {
  324. var output bytes.Buffer
  325. err := tp.Execute(&output, r)
  326. if err != nil {
  327. logger.Warnf("sink node %s instance %d publish %s decode template error: %v", ctx.GetOpId(), ctx.GetInstanceId(), val, err)
  328. stats.IncTotalExceptions()
  329. return nil
  330. }
  331. outdatas = append(outdatas, output.Bytes())
  332. } else {
  333. if ot, e := json.Marshal(r); e != nil {
  334. logger.Warnf("sink node %s instance %d publish %s marshal error: %v", ctx.GetOpId(), ctx.GetInstanceId(), r, e)
  335. stats.IncTotalExceptions()
  336. return nil
  337. } else {
  338. outdatas = append(outdatas, ot)
  339. }
  340. }
  341. }
  342. }
  343. case error:
  344. outdatas = [][]byte{[]byte(fmt.Sprintf(`[{"error":"%s"}]`, val.Error()))}
  345. default:
  346. outdatas = [][]byte{[]byte(fmt.Sprintf(`[{"error":"result is not a string but found %#v"}]`, val))}
  347. }
  348. return outdatas
  349. }
  350. func doCollectCacheTuple(sink api.Sink, item *CacheTuple, stats StatManager, retryInterval, retryCount int, omitIfEmpty bool, sendSingle bool, tp *template.Template, signalCh chan<- int, ctx api.StreamContext) {
  351. stats.IncTotalRecordsIn()
  352. stats.ProcessTimeStart()
  353. defer stats.ProcessTimeEnd()
  354. logger := ctx.GetLogger()
  355. outdatas := getOutData(stats, ctx, item.data, omitIfEmpty, sendSingle, tp)
  356. for _, outdata := range outdatas {
  357. outerloop:
  358. for {
  359. select {
  360. case <-ctx.Done():
  361. logger.Infof("sink node %s instance %d stops data resending", ctx.GetOpId(), ctx.GetInstanceId())
  362. return
  363. default:
  364. if err := sink.Collect(ctx, outdata); err != nil {
  365. stats.IncTotalExceptions()
  366. logger.Warnf("sink node %s instance %d publish %s error: %v", ctx.GetOpId(), ctx.GetInstanceId(), outdata, err)
  367. if retryInterval > 0 && retryCount > 0 {
  368. retryCount--
  369. time.Sleep(time.Duration(retryInterval) * time.Millisecond)
  370. logger.Debugf("try again")
  371. } else {
  372. break outerloop
  373. }
  374. } else {
  375. logger.Debugf("success")
  376. stats.IncTotalRecordsOut()
  377. select {
  378. case signalCh <- item.index:
  379. default:
  380. logger.Warnf("sink cache missing response for %d", item.index)
  381. }
  382. break outerloop
  383. }
  384. }
  385. }
  386. }
  387. }
  388. func getSink(name string, action map[string]interface{}) (api.Sink, error) {
  389. var (
  390. s api.Sink
  391. err error
  392. )
  393. s, err = io.Sink(name)
  394. if s != nil {
  395. err = s.Configure(action)
  396. if err != nil {
  397. return nil, err
  398. }
  399. return s, nil
  400. } else {
  401. if err != nil {
  402. return nil, err
  403. } else {
  404. return nil, fmt.Errorf("sink %s not found", name)
  405. }
  406. }
  407. }
  408. //Override defaultNode
  409. func (m *SinkNode) AddOutput(_ chan<- interface{}, name string) error {
  410. return fmt.Errorf("fail to add output %s, sink %s cannot add output", name, m.name)
  411. }
  412. //Override defaultNode
  413. func (m *SinkNode) Broadcast(_ interface{}) error {
  414. return fmt.Errorf("sink %s cannot add broadcast", m.name)
  415. }
  416. func (m *SinkNode) drainError(errCh chan<- error, err error, ctx api.StreamContext, logger api.Logger) {
  417. go func() {
  418. select {
  419. case errCh <- err:
  420. ctx.GetLogger().Errorf("error in sink %s", err)
  421. case <-ctx.Done():
  422. m.close(ctx, logger)
  423. }
  424. }()
  425. }
  426. func (m *SinkNode) close(ctx api.StreamContext, logger api.Logger) {
  427. for _, s := range m.sinks {
  428. if err := s.Close(ctx); err != nil {
  429. logger.Warnf("close sink fails: %v", err)
  430. }
  431. }
  432. if m.tch != nil {
  433. close(m.tch)
  434. m.tch = nil
  435. }
  436. }
  437. // Only called when checkpoint enabled
  438. func (m *SinkNode) SaveCache() {
  439. m.tch <- struct{}{}
  440. }