123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510 |
- package kinesisacquisition
- import (
- "bytes"
- "compress/gzip"
- "encoding/json"
- "fmt"
- "io/ioutil"
- "strings"
- "time"
- "github.com/aws/aws-sdk-go/aws"
- "github.com/aws/aws-sdk-go/aws/arn"
- "github.com/aws/aws-sdk-go/aws/session"
- "github.com/aws/aws-sdk-go/service/kinesis"
- "github.com/crowdsecurity/crowdsec/pkg/acquisition/configuration"
- "github.com/crowdsecurity/crowdsec/pkg/leakybucket"
- "github.com/crowdsecurity/crowdsec/pkg/types"
- "github.com/pkg/errors"
- "github.com/prometheus/client_golang/prometheus"
- log "github.com/sirupsen/logrus"
- "gopkg.in/tomb.v2"
- "gopkg.in/yaml.v2"
- )
- type KinesisConfiguration struct {
- configuration.DataSourceCommonCfg `yaml:",inline"`
- StreamName string `yaml:"stream_name"`
- StreamARN string `yaml:"stream_arn"`
- UseEnhancedFanOut bool `yaml:"use_enhanced_fanout"` //Use RegisterStreamConsumer and SubscribeToShard instead of GetRecords
- AwsProfile *string `yaml:"aws_profile"`
- AwsRegion string `yaml:"aws_region"`
- AwsEndpoint string `yaml:"aws_endpoint"`
- ConsumerName string `yaml:"consumer_name"`
- FromSubscription bool `yaml:"from_subscription"`
- MaxRetries int `yaml:"max_retries"`
- }
- type KinesisSource struct {
- Config KinesisConfiguration
- logger *log.Entry
- kClient *kinesis.Kinesis
- shardReaderTomb *tomb.Tomb
- }
- type CloudWatchSubscriptionRecord struct {
- MessageType string `json:"messageType"`
- Owner string `json:"owner"`
- LogGroup string `json:"logGroup"`
- LogStream string `json:"logStream"`
- SubscriptionFilters []string `json:"subscriptionFilters"`
- LogEvents []CloudwatchSubscriptionLogEvent `json:"logEvents"`
- }
- type CloudwatchSubscriptionLogEvent struct {
- ID string `json:"id"`
- Message string `json:"message"`
- Timestamp int64 `json:"timestamp"`
- }
- var linesRead = prometheus.NewCounterVec(
- prometheus.CounterOpts{
- Name: "cs_kinesis_stream_hits_total",
- Help: "Number of event read per stream.",
- },
- []string{"stream"},
- )
- var linesReadShards = prometheus.NewCounterVec(
- prometheus.CounterOpts{
- Name: "cs_kinesis_shards_hits_total",
- Help: "Number of event read per shards.",
- },
- []string{"stream", "shard"},
- )
- func (k *KinesisSource) newClient() error {
- var sess *session.Session
- if k.Config.AwsProfile != nil {
- sess = session.Must(session.NewSessionWithOptions(session.Options{
- SharedConfigState: session.SharedConfigEnable,
- Profile: *k.Config.AwsProfile,
- }))
- } else {
- sess = session.Must(session.NewSessionWithOptions(session.Options{
- SharedConfigState: session.SharedConfigEnable,
- }))
- }
- if sess == nil {
- return fmt.Errorf("failed to create aws session")
- }
- config := aws.NewConfig()
- if k.Config.AwsRegion != "" {
- config = config.WithRegion(k.Config.AwsRegion)
- }
- if k.Config.AwsEndpoint != "" {
- config = config.WithEndpoint(k.Config.AwsEndpoint)
- }
- k.kClient = kinesis.New(sess, config)
- if k.kClient == nil {
- return fmt.Errorf("failed to create kinesis client")
- }
- return nil
- }
- func (k *KinesisSource) GetMetrics() []prometheus.Collector {
- return []prometheus.Collector{linesRead, linesReadShards}
- }
- func (k *KinesisSource) GetAggregMetrics() []prometheus.Collector {
- return []prometheus.Collector{linesRead, linesReadShards}
- }
- func (k *KinesisSource) Configure(yamlConfig []byte, logger *log.Entry) error {
- config := KinesisConfiguration{}
- k.logger = logger
- err := yaml.UnmarshalStrict(yamlConfig, &config)
- if err != nil {
- return errors.Wrap(err, "Cannot parse kinesis datasource configuration")
- }
- if config.Mode == "" {
- config.Mode = configuration.TAIL_MODE
- }
- k.Config = config
- if k.Config.StreamName == "" && !k.Config.UseEnhancedFanOut {
- return fmt.Errorf("stream_name is mandatory when use_enhanced_fanout is false")
- }
- if k.Config.StreamARN == "" && k.Config.UseEnhancedFanOut {
- return fmt.Errorf("stream_arn is mandatory when use_enhanced_fanout is true")
- }
- if k.Config.ConsumerName == "" && k.Config.UseEnhancedFanOut {
- return fmt.Errorf("consumer_name is mandatory when use_enhanced_fanout is true")
- }
- if k.Config.StreamARN != "" && k.Config.StreamName != "" {
- return fmt.Errorf("stream_arn and stream_name are mutually exclusive")
- }
- if k.Config.MaxRetries <= 0 {
- k.Config.MaxRetries = 10
- }
- err = k.newClient()
- if err != nil {
- return errors.Wrap(err, "Cannot create kinesis client")
- }
- k.shardReaderTomb = &tomb.Tomb{}
- return nil
- }
- func (k *KinesisSource) ConfigureByDSN(string, map[string]string, *log.Entry) error {
- return fmt.Errorf("kinesis datasource does not support command-line acquisition")
- }
- func (k *KinesisSource) GetMode() string {
- return k.Config.Mode
- }
- func (k *KinesisSource) GetName() string {
- return "kinesis"
- }
- func (k *KinesisSource) OneShotAcquisition(out chan types.Event, t *tomb.Tomb) error {
- return fmt.Errorf("kinesis datasource does not support one-shot acquisition")
- }
- func (k *KinesisSource) decodeFromSubscription(record []byte) ([]CloudwatchSubscriptionLogEvent, error) {
- b := bytes.NewBuffer(record)
- r, err := gzip.NewReader(b)
- if err != nil {
- k.logger.Error(err)
- return nil, err
- }
- decompressed, err := ioutil.ReadAll(r)
- if err != nil {
- k.logger.Error(err)
- return nil, err
- }
- var subscriptionRecord CloudWatchSubscriptionRecord
- err = json.Unmarshal(decompressed, &subscriptionRecord)
- if err != nil {
- k.logger.Error(err)
- return nil, err
- }
- return subscriptionRecord.LogEvents, nil
- }
- func (k *KinesisSource) WaitForConsumerDeregistration(consumerName string, streamARN string) error {
- maxTries := k.Config.MaxRetries
- for i := 0; i < maxTries; i++ {
- _, err := k.kClient.DescribeStreamConsumer(&kinesis.DescribeStreamConsumerInput{
- ConsumerName: aws.String(consumerName),
- StreamARN: aws.String(streamARN),
- })
- if err != nil {
- switch err.(type) {
- case *kinesis.ResourceNotFoundException:
- return nil
- default:
- k.logger.Errorf("Error while waiting for consumer deregistration: %s", err)
- return errors.Wrap(err, "Cannot describe stream consumer")
- }
- }
- time.Sleep(time.Millisecond * 200 * time.Duration(i+1))
- }
- return fmt.Errorf("consumer %s is not deregistered after %d tries", consumerName, maxTries)
- }
- func (k *KinesisSource) DeregisterConsumer() error {
- k.logger.Debugf("Deregistering consumer %s if it exists", k.Config.ConsumerName)
- _, err := k.kClient.DeregisterStreamConsumer(&kinesis.DeregisterStreamConsumerInput{
- ConsumerName: aws.String(k.Config.ConsumerName),
- StreamARN: aws.String(k.Config.StreamARN),
- })
- if err != nil {
- switch err.(type) {
- case *kinesis.ResourceNotFoundException:
- default:
- return errors.Wrap(err, "Cannot deregister stream consumer")
- }
- }
- err = k.WaitForConsumerDeregistration(k.Config.ConsumerName, k.Config.StreamARN)
- if err != nil {
- return errors.Wrap(err, "Cannot wait for consumer deregistration")
- }
- return nil
- }
- func (k *KinesisSource) WaitForConsumerRegistration(consumerARN string) error {
- maxTries := k.Config.MaxRetries
- for i := 0; i < maxTries; i++ {
- describeOutput, err := k.kClient.DescribeStreamConsumer(&kinesis.DescribeStreamConsumerInput{
- ConsumerARN: aws.String(consumerARN),
- })
- if err != nil {
- return errors.Wrap(err, "Cannot describe stream consumer")
- }
- if *describeOutput.ConsumerDescription.ConsumerStatus == "ACTIVE" {
- k.logger.Debugf("Consumer %s is active", consumerARN)
- return nil
- }
- time.Sleep(time.Millisecond * 200 * time.Duration(i+1))
- k.logger.Debugf("Waiting for consumer registration %d", i)
- }
- return fmt.Errorf("consumer %s is not active after %d tries", consumerARN, maxTries)
- }
- func (k *KinesisSource) RegisterConsumer() (*kinesis.RegisterStreamConsumerOutput, error) {
- k.logger.Debugf("Registering consumer %s", k.Config.ConsumerName)
- streamConsumer, err := k.kClient.RegisterStreamConsumer(&kinesis.RegisterStreamConsumerInput{
- ConsumerName: aws.String(k.Config.ConsumerName),
- StreamARN: aws.String(k.Config.StreamARN),
- })
- if err != nil {
- return nil, errors.Wrap(err, "Cannot register stream consumer")
- }
- err = k.WaitForConsumerRegistration(*streamConsumer.Consumer.ConsumerARN)
- if err != nil {
- return nil, errors.Wrap(err, "Timeout while waiting for consumer to be active")
- }
- return streamConsumer, nil
- }
- func (k *KinesisSource) ParseAndPushRecords(records []*kinesis.Record, out chan types.Event, logger *log.Entry, shardId string) {
- for _, record := range records {
- if k.Config.StreamARN != "" {
- linesReadShards.With(prometheus.Labels{"stream": k.Config.StreamARN, "shard": shardId}).Inc()
- linesRead.With(prometheus.Labels{"stream": k.Config.StreamARN}).Inc()
- } else {
- linesReadShards.With(prometheus.Labels{"stream": k.Config.StreamName, "shard": shardId}).Inc()
- linesRead.With(prometheus.Labels{"stream": k.Config.StreamName}).Inc()
- }
- var data []CloudwatchSubscriptionLogEvent
- var err error
- if k.Config.FromSubscription {
- //The AWS docs says that the data is base64 encoded
- //but apparently GetRecords decodes it for us ?
- data, err = k.decodeFromSubscription(record.Data)
- if err != nil {
- logger.Errorf("Cannot decode data: %s", err)
- continue
- }
- } else {
- data = []CloudwatchSubscriptionLogEvent{{Message: string(record.Data)}}
- }
- for _, event := range data {
- logger.Tracef("got record %s", event.Message)
- l := types.Line{}
- l.Raw = event.Message
- l.Labels = k.Config.Labels
- l.Time = time.Now().UTC()
- l.Process = true
- l.Module = k.GetName()
- if k.Config.StreamARN != "" {
- l.Src = k.Config.StreamARN
- } else {
- l.Src = k.Config.StreamName
- }
- evt := types.Event{Line: l, Process: true, Type: types.LOG, ExpectMode: leakybucket.LIVE}
- out <- evt
- }
- }
- }
- func (k *KinesisSource) ReadFromSubscription(reader kinesis.SubscribeToShardEventStreamReader, out chan types.Event, shardId string, streamName string) error {
- logger := k.logger.WithFields(log.Fields{"shard_id": shardId})
- //ghetto sync, kinesis allows to subscribe to a closed shard, which will make the goroutine exit immediately
- //and we won't be able to start a new one if this is the first one started by the tomb
- //TODO: look into parent shards to see if a shard is closed before starting to read it ?
- time.Sleep(time.Second)
- for {
- select {
- case <-k.shardReaderTomb.Dying():
- logger.Infof("Subscribed shard reader is dying")
- err := reader.Close()
- if err != nil {
- return errors.Wrap(err, "Cannot close kinesis subscribed shard reader")
- }
- return nil
- case event, ok := <-reader.Events():
- if !ok {
- logger.Infof("Event chan has been closed")
- return nil
- }
- switch event := event.(type) {
- case *kinesis.SubscribeToShardEvent:
- k.ParseAndPushRecords(event.Records, out, logger, shardId)
- case *kinesis.SubscribeToShardEventStreamUnknownEvent:
- logger.Infof("got an unknown event, what to do ?")
- }
- }
- }
- }
- func (k *KinesisSource) SubscribeToShards(arn arn.ARN, streamConsumer *kinesis.RegisterStreamConsumerOutput, out chan types.Event) error {
- shards, err := k.kClient.ListShards(&kinesis.ListShardsInput{
- StreamName: aws.String(arn.Resource[7:]),
- })
- if err != nil {
- return errors.Wrap(err, "Cannot list shards for enhanced_read")
- }
- for _, shard := range shards.Shards {
- shardId := *shard.ShardId
- r, err := k.kClient.SubscribeToShard(&kinesis.SubscribeToShardInput{
- ShardId: aws.String(shardId),
- StartingPosition: &kinesis.StartingPosition{Type: aws.String(kinesis.ShardIteratorTypeLatest)},
- ConsumerARN: streamConsumer.Consumer.ConsumerARN,
- })
- if err != nil {
- return errors.Wrap(err, "Cannot subscribe to shard")
- }
- k.shardReaderTomb.Go(func() error {
- return k.ReadFromSubscription(r.GetEventStream().Reader, out, shardId, arn.Resource[7:])
- })
- }
- return nil
- }
- func (k *KinesisSource) EnhancedRead(out chan types.Event, t *tomb.Tomb) error {
- parsedARN, err := arn.Parse(k.Config.StreamARN)
- if err != nil {
- return errors.Wrap(err, "Cannot parse stream ARN")
- }
- if !strings.HasPrefix(parsedARN.Resource, "stream/") {
- return fmt.Errorf("resource part of stream ARN %s does not start with stream/", k.Config.StreamARN)
- }
- k.logger = k.logger.WithFields(log.Fields{"stream": parsedARN.Resource[7:]})
- k.logger.Info("starting kinesis acquisition with enhanced fan-out")
- err = k.DeregisterConsumer()
- if err != nil {
- return errors.Wrap(err, "Cannot deregister consumer")
- }
- streamConsumer, err := k.RegisterConsumer()
- if err != nil {
- return errors.Wrap(err, "Cannot register consumer")
- }
- for {
- k.shardReaderTomb = &tomb.Tomb{}
- err = k.SubscribeToShards(parsedARN, streamConsumer, out)
- if err != nil {
- return errors.Wrap(err, "Cannot subscribe to shards")
- }
- select {
- case <-t.Dying():
- k.logger.Infof("Kinesis source is dying")
- k.shardReaderTomb.Kill(nil)
- _ = k.shardReaderTomb.Wait() //we don't care about the error as we kill the tomb ourselves
- err = k.DeregisterConsumer()
- if err != nil {
- return errors.Wrap(err, "Cannot deregister consumer")
- }
- return nil
- case <-k.shardReaderTomb.Dying():
- k.logger.Debugf("Kinesis subscribed shard reader is dying")
- if k.shardReaderTomb.Err() != nil {
- return k.shardReaderTomb.Err()
- }
- //All goroutines have exited without error, so a resharding event, start again
- k.logger.Debugf("All reader goroutines have exited, resharding event or periodic resubscribe")
- continue
- }
- }
- }
- func (k *KinesisSource) ReadFromShard(out chan types.Event, shardId string) error {
- logger := k.logger.WithFields(log.Fields{"shard": shardId})
- logger.Debugf("Starting to read shard")
- sharIt, err := k.kClient.GetShardIterator(&kinesis.GetShardIteratorInput{ShardId: aws.String(shardId),
- StreamName: &k.Config.StreamName,
- ShardIteratorType: aws.String(kinesis.ShardIteratorTypeLatest)})
- if err != nil {
- logger.Errorf("Cannot get shard iterator: %s", err)
- return errors.Wrap(err, "Cannot get shard iterator")
- }
- it := sharIt.ShardIterator
- //AWS recommends to wait for a second between calls to GetRecords for a given shard
- ticker := time.NewTicker(time.Second)
- for {
- select {
- case <-ticker.C:
- records, err := k.kClient.GetRecords(&kinesis.GetRecordsInput{ShardIterator: it})
- it = records.NextShardIterator
- if err != nil {
- switch err.(type) {
- case *kinesis.ProvisionedThroughputExceededException:
- logger.Warn("Provisioned throughput exceeded")
- //TODO: implement exponential backoff
- continue
- case *kinesis.ExpiredIteratorException:
- logger.Warn("Expired iterator")
- continue
- default:
- logger.Error("Cannot get records")
- return errors.Wrap(err, "Cannot get records")
- }
- }
- k.ParseAndPushRecords(records.Records, out, logger, shardId)
- if it == nil {
- logger.Warnf("Shard has been closed")
- return nil
- }
- case <-k.shardReaderTomb.Dying():
- logger.Infof("shardReaderTomb is dying, exiting ReadFromShard")
- ticker.Stop()
- return nil
- }
- }
- }
- func (k *KinesisSource) ReadFromStream(out chan types.Event, t *tomb.Tomb) error {
- k.logger = k.logger.WithFields(log.Fields{"stream": k.Config.StreamName})
- k.logger.Info("starting kinesis acquisition from shards")
- for {
- shards, err := k.kClient.ListShards(&kinesis.ListShardsInput{
- StreamName: aws.String(k.Config.StreamName),
- })
- if err != nil {
- return errors.Wrap(err, "Cannot list shards")
- }
- k.shardReaderTomb = &tomb.Tomb{}
- for _, shard := range shards.Shards {
- shardId := *shard.ShardId
- k.shardReaderTomb.Go(func() error {
- defer types.CatchPanic("crowdsec/acquis/kinesis/streaming/shard")
- return k.ReadFromShard(out, shardId)
- })
- }
- select {
- case <-t.Dying():
- k.logger.Info("kinesis source is dying")
- k.shardReaderTomb.Kill(nil)
- _ = k.shardReaderTomb.Wait() //we don't care about the error as we kill the tomb ourselves
- return nil
- case <-k.shardReaderTomb.Dying():
- reason := k.shardReaderTomb.Err()
- if reason != nil {
- k.logger.Errorf("Unexpected error from shard reader : %s", reason)
- return reason
- }
- k.logger.Infof("All shards have been closed, probably a resharding event, restarting acquisition")
- continue
- }
- }
- }
- func (k *KinesisSource) StreamingAcquisition(out chan types.Event, t *tomb.Tomb) error {
- t.Go(func() error {
- defer types.CatchPanic("crowdsec/acquis/kinesis/streaming")
- if k.Config.UseEnhancedFanOut {
- return k.EnhancedRead(out, t)
- } else {
- return k.ReadFromStream(out, t)
- }
- })
- return nil
- }
- func (k *KinesisSource) CanRun() error {
- return nil
- }
- func (k *KinesisSource) Dump() interface{} {
- return k
- }
|