moby/vendor/cloud.google.com/go
Akihiro Suda 5a1b06d7fd rerun vndr
* run latest vndr so as to collect more LICENSE files
 * remove unused packages
 * vendor github.com/philhofer/fwd with LICENSE.md (MIT)
 * vendor github.com/bsphere/le_go with LICENSE (MIT)

Signed-off-by: Akihiro Suda <suda.akihiro@lab.ntt.co.jp>
2017-03-08 02:29:34 +00:00
..
compute/metadata vendor oauth2@96382aa0 and add its deps 2016-12-16 04:37:47 +00:00
internal vendor oauth2@96382aa0 and add its deps 2016-12-16 04:37:47 +00:00
logging rerun vndr 2017-03-08 02:29:34 +00:00
LICENSE vendor oauth2@96382aa0 and add its deps 2016-12-16 04:37:47 +00:00
README.md rerun vndr 2017-03-08 02:29:34 +00:00

Google Cloud for Go

Build Status GoDoc

import "cloud.google.com/go"

Go packages for Google Cloud Platform services.

To install the packages on your system,

$ go get -u cloud.google.com/go/...

NOTE: These packages are under development, and may occasionally make backwards-incompatible changes.

NOTE: Github repo is a mirror of https://code.googlesource.com/gocloud.

News

December 5, 2016

More changes to BigQuery:

  • The ValueList type was removed. It is no longer necessary. Instead of

    var v ValueList
    ... it.Next(&v) ..
    

    use

    var v []Value
    ... it.Next(&v) ...
    
  • Previously, repeatedly calling RowIterator.Next on the same []Value or ValueList would append to the slice. Now each call resets the size to zero first.

  • Schema inference will infer the SQL type BYTES for a struct field of type []byte. Previously it inferred STRING.

  • The types uint, uint64 and uintptr are no longer supported in schema inference. BigQuery's integer type is INT64, and those types may hold values that are not correctly represented in a 64-bit signed integer.

  • The SQL types DATE, TIME and DATETIME are now supported. They correspond to the Date, Time and DateTime types in the new cloud.google.com/go/civil package.

November 17, 2016

Change to BigQuery: values from INTEGER columns will now be returned as int64, not int. This will avoid errors arising from large values on 32-bit systems.

November 8, 2016

New datastore feature: datastore now encodes your nested Go structs as Entity values, instead of a flattened list of the embedded struct's fields. This means that you may now have twice-nested slices, eg.

type State struct {
  Cities  []struct{
    Populations []int
  }
}

See the announcement for more details.

November 8, 2016

Breaking changes to datastore: contexts no longer hold namespaces; instead you must set a key's namespace explicitly. Also, key functions have been changed and renamed.

  • The WithNamespace function has been removed. To specify a namespace in a Query, use the Query.Namespace method:

    q := datastore.NewQuery("Kind").Namespace("ns")
    
  • All the fields of Key are exported. That means you can construct any Key with a struct literal:

    k := &Key{Kind: "Kind",  ID: 37, Namespace: "ns"}
    
  • As a result of the above, the Key methods Kind, ID, d.Name, Parent, SetParent and Namespace have been removed.

  • NewIncompleteKey has been removed, replaced by IncompleteKey. Replace

    NewIncompleteKey(ctx, kind, parent)
    

    with

    IncompleteKey(kind, parent)
    

    and if you do use namespaces, make sure you set the namespace on the returned key.

  • NewKey has been removed, replaced by NameKey and IDKey. Replace

    NewKey(ctx, kind, name, 0, parent)
    NewKey(ctx, kind, "", id, parent)
    

    with

    NameKey(kind, name, parent)
    IDKey(kind, id, parent)
    

    and if you do use namespaces, make sure you set the namespace on the returned key.

  • The Done variable has been removed. Replace datastore.Done with iterator.Done, from the package google.golang.org/api/iterator.

  • The Client.Close method will have a return type of error. It will return the result of closing the underlying gRPC connection.

See the announcement for more details.

October 27, 2016

Breaking change to bigquery: NewGCSReference is now a function, not a method on Client.

New bigquery feature: Table.LoaderFrom now accepts a ReaderSource, enabling loading data into a table from a file or any io.Reader.

October 21, 2016

Breaking change to pubsub: removed pubsub.Done.

Use iterator.Done instead, where iterator is the package google.golang.org/api/iterator.

Older news

Supported APIs

Google API Status Package
Datastore beta cloud.google.com/go/datastore
Storage beta cloud.google.com/go/storage
Bigtable beta cloud.google.com/go/bigtable
BigQuery beta cloud.google.com/go/bigquery
Logging beta cloud.google.com/go/logging
Pub/Sub experimental cloud.google.com/go/pubsub
Vision experimental cloud.google.com/go/vision
Language experimental cloud.google.com/go/language/apiv1beta1
Speech experimental cloud.google.com/go/speech/apiv1beta

Experimental status: the API is still being actively developed. As a result, it might change in backward-incompatible ways and is not recommended for production use.

Beta status: the API is largely complete, but still has outstanding features and bugs to be addressed. There may be minor backwards-incompatible changes where necessary.

Stable status: the API is mature and ready for production use. We will continue addressing bugs and feature requests.

Documentation and examples are available at https://godoc.org/cloud.google.com/go

Visit or join the google-api-go-announce group for updates on these packages.

Go Versions Supported

We support the two most recent major versions of Go. If Google App Engine uses an older version, we support that as well. You can see which versions are currently supported by looking at the lines following go: in .travis.yml.

Authorization

By default, each API will use Google Application Default Credentials for authorization credentials used in calling the API endpoints. This will allow your application to run in many environments without requiring explicit configuration.

To authorize using a JSON key file, pass option.WithServiceAccountFile to the NewClient function of the desired package. For example:

client, err := storage.NewClient(ctx, option.WithServiceAccountFile("path/to/keyfile.json"))

You can exert more control over authorization by using the golang.org/x/oauth2 package to create an oauth2.TokenSource. Then pass option.WithTokenSource to the NewClient function:

tokenSource := ...
client, err := storage.NewClient(ctx, option.WithTokenSource(tokenSource))

Cloud Datastore GoDoc

Example Usage

First create a datastore.Client to use throughout your application:

client, err := datastore.NewClient(ctx, "my-project-id")
if err != nil {
	log.Fatal(err)
}

Then use that client to interact with the API:

type Post struct {
	Title       string
	Body        string `datastore:",noindex"`
	PublishedAt time.Time
}
keys := []*datastore.Key{
	datastore.NewKey(ctx, "Post", "post1", 0, nil),
	datastore.NewKey(ctx, "Post", "post2", 0, nil),
}
posts := []*Post{
	{Title: "Post 1", Body: "...", PublishedAt: time.Now()},
	{Title: "Post 2", Body: "...", PublishedAt: time.Now()},
}
if _, err := client.PutMulti(ctx, keys, posts); err != nil {
	log.Fatal(err)
}

Cloud Storage GoDoc

Example Usage

First create a storage.Client to use throughout your application:

client, err := storage.NewClient(ctx)
if err != nil {
	log.Fatal(err)
}
// Read the object1 from bucket.
rc, err := client.Bucket("bucket").Object("object1").NewReader(ctx)
if err != nil {
	log.Fatal(err)
}
defer rc.Close()
body, err := ioutil.ReadAll(rc)
if err != nil {
	log.Fatal(err)
}

Cloud Pub/Sub GoDoc

Example Usage

First create a pubsub.Client to use throughout your application:

client, err := pubsub.NewClient(ctx, "project-id")
if err != nil {
	log.Fatal(err)
}
// Publish "hello world" on topic1.
topic := client.Topic("topic1")
msgIDs, err := topic.Publish(ctx, &pubsub.Message{
	Data: []byte("hello world"),
})
if err != nil {
	log.Fatal(err)
}

// Create an iterator to pull messages via subscription1.
it, err := client.Subscription("subscription1").Pull(ctx)
if err != nil {
	log.Println(err)
}
defer it.Stop()

// Consume N messages from the iterator.
for i := 0; i < N; i++ {
	msg, err := it.Next()
	if err == iterator.Done {
		break
	}
	if err != nil {
		log.Fatalf("Failed to retrieve message: %v", err)
	}

	fmt.Printf("Message %d: %s\n", i, msg.Data)
	msg.Done(true) // Acknowledge that we've consumed the message.
}

Cloud BigQuery GoDoc

Example Usage

First create a bigquery.Client to use throughout your application:

c, err := bigquery.NewClient(ctx, "my-project-ID")
if err != nil {
    // TODO: Handle error.
}

Then use that client to interact with the API:

// Construct a query.
q := c.Query(`
    SELECT year, SUM(number)
    FROM [bigquery-public-data:usa_names.usa_1910_2013]
    WHERE name = "William"
    GROUP BY year
    ORDER BY year
`)
// Execute the query.
it, err := q.Read(ctx)
if err != nil {
    // TODO: Handle error.
}
// Iterate through the results.
for {
    var values bigquery.ValueList
    err := it.Next(&values)
    if err == iterator.Done {
        break
    }
    if err != nil {
        // TODO: Handle error.
    }
    fmt.Println(values)
}

Stackdriver Logging GoDoc

Example Usage

First create a logging.Client to use throughout your application:

ctx := context.Background()
client, err := logging.NewClient(ctx, "my-project")
if err != nil {
    // TODO: Handle error.
}

Usually, you'll want to add log entries to a buffer to be periodically flushed (automatically and asynchronously) to the Stackdriver Logging service.

logger := client.Logger("my-log")
logger.Log(logging.Entry{Payload: "something happened!"})

Close your client before your program exits, to flush any buffered log entries.

err = client.Close()
if err != nil {
    // TODO: Handle error.
}

Contributing

Contributions are welcome. Please, see the CONTRIBUTING document for details. We're using Gerrit for our code reviews. Please don't open pull requests against this repo, new pull requests will be automatically closed.

Please note that this project is released with a Contributor Code of Conduct. By participating in this project you agree to abide by its terms. See Contributor Code of Conduct for more information.