import "cloud.google.com/go"
Go packages for Google Cloud Platform services.
NOTE: These packages are under development, and may occasionally make backwards-incompatible changes.
NOTE: Github repo is a mirror of https://code.googlesource.com/gocloud.
October 10, 2016
Breaking changes to cloud.google.com/go/storage:
AdminClient replaced by methods on Client. Replace
adminClient.CreateBucket(ctx, bucketName, attrs)
with
client.Bucket(bucketName).Create(ctx, projectID, attrs)
BucketHandle.List replaced by BucketHandle.Objects. Replace
for query != nil {
objs, err := bucket.List(d.ctx, query)
if err != nil { ... }
query = objs.Next
for _, obj := range objs.Results {
fmt.Println(obj)
}
}
with
iter := bucket.Objects(d.ctx, query)
for {
obj, err := iter.Next()
if err == iterator.Done {
break
}
if err != nil { ... }
fmt.Println(obj)
}
(The iterator
package is at google.golang.org/api/iterator
.)
Replace Query.Cursor
with ObjectIterator.PageInfo().Token
.
Replace Query.MaxResults
with ObjectIterator.PageInfo().MaxSize
.
ObjectHandle.CopyTo replaced by ObjectHandle.CopierFrom. Replace
attrs, err := src.CopyTo(ctx, dst, nil)
with
attrs, err := dst.CopierFrom(src).Run(ctx)
Replace
attrs, err := src.CopyTo(ctx, dst, &storage.ObjectAttrs{ContextType: "text/html"})
with
c := dst.CopierFrom(src)
c.ContextType = "text/html"
attrs, err := c.Run(ctx)
ObjectHandle.ComposeFrom replaced by ObjectHandle.ComposerFrom. Replace
attrs, err := dst.ComposeFrom(ctx, []*storage.ObjectHandle{src1, src2}, nil)
with
attrs, err := dst.ComposerFrom(src1, src2).Run(ctx)
ObjectHandle.Update's ObjectAttrs argument replaced by ObjectAttrsToUpdate. Replace
attrs, err := obj.Update(ctx, &storage.ObjectAttrs{ContextType: "text/html"})
with
attrs, err := obj.Update(ctx, storage.ObjectAttrsToUpdate{ContextType: "text/html"})
ObjectHandle.WithConditions replaced by ObjectHandle.If. Replace
obj.WithConditions(storage.Generation(gen), storage.IfMetaGenerationMatch(mgen))
with
obj.Generation(gen).If(storage.Conditions{MetagenerationMatch: mgen})
Replace
obj.WithConditions(storage.IfGenerationMatch(0))
with
obj.If(storage.Conditions{DoesNotExist: true})
storage.Done
replaced by iterator.Done
(from package google.golang.org/api/iterator
).
October 6, 2016
Package preview/logging deleted. Use logging instead.
September 27, 2016
Logging client replaced with preview version (see below).
September 8, 2016
New clients for some of Google's Machine Learning APIs: Vision, Speech, and Natural Language.
Preview version of a new Stackdriver Logging client in cloud.google.com/go/preview/logging
. This client uses gRPC as its transport layer, and supports log reading, sinks and metrics. It will replace the current client at cloud.google.com/go/logging
shortly.
Google API | Status | Package |
---|---|---|
Datastore | beta | cloud.google.com/go/datastore |
Storage | beta | cloud.google.com/go/storage |
Pub/Sub | experimental | cloud.google.com/go/pubsub |
Bigtable | beta | cloud.google.com/go/bigtable |
BigQuery | experimental | cloud.google.com/go/bigquery |
Logging | experimental | cloud.google.com/go/logging |
Vision | experimental | cloud.google.com/go/vision |
Language | experimental | cloud.google.com/go/language/apiv1beta1 |
Speech | experimental | cloud.google.com/go/speech/apiv1beta |
Experimental status: the API is still being actively developed. As a result, it might change in backward-incompatible ways and is not recommended for production use.
Beta status: the API is largely complete, but still has outstanding features and bugs to be addressed. There may be minor backwards-incompatible changes where necessary.
Stable status: the API is mature and ready for production use. We will continue addressing bugs and feature requests.
Documentation and examples are available at https://godoc.org/cloud.google.com/go
Visit or join the google-api-go-announce group for updates on these packages.
We support the two most recent major versions of Go. If Google App Engine uses an older version, we support that as well. You can see which versions are currently supported by looking at the lines following go:
in .travis.yml
.
By default, each API will use Google Application Default Credentials for authorization credentials used in calling the API endpoints. This will allow your application to run in many environments without requiring explicit configuration.
To authorize using a JSON key file, pass option.WithServiceAccountFile
to the NewClient
function of the desired package. For example:
client, err := storage.NewClient(ctx, option.WithServiceAccountFile("path/to/keyfile.json"))
You can exert more control over authorization by using the golang.org/x/oauth2
package to create an oauth2.TokenSource
. Then pass option.WithTokenSource
to the NewClient
function:
tokenSource := ...
client, err := storage.NewClient(ctx, option.WithTokenSource(tokenSource))
First create a datastore.Client
to use throughout your application:
client, err := datastore.NewClient(ctx, "my-project-id")
if err != nil {
log.Fatal(err)
}
Then use that client to interact with the API:
type Post struct {
Title string
Body string `datastore:",noindex"`
PublishedAt time.Time
}
keys := []*datastore.Key{
datastore.NewKey(ctx, "Post", "post1", 0, nil),
datastore.NewKey(ctx, "Post", "post2", 0, nil),
}
posts := []*Post{
{Title: "Post 1", Body: "...", PublishedAt: time.Now()},
{Title: "Post 2", Body: "...", PublishedAt: time.Now()},
}
if _, err := client.PutMulti(ctx, keys, posts); err != nil {
log.Fatal(err)
}
First create a storage.Client
to use throughout your application:
client, err := storage.NewClient(ctx)
if err != nil {
log.Fatal(err)
}
// Read the object1 from bucket. rc, err := client.Bucket("bucket").Object("object1").NewReader(ctx) if err != nil { log.Fatal(err) } defer rc.Close() body, err := ioutil.ReadAll(rc) if err != nil { log.Fatal(err) }
First create a pubsub.Client
to use throughout your application:
client, err := pubsub.NewClient(ctx, "project-id")
if err != nil {
log.Fatal(err)
}
// Publish "hello world" on topic1. topic := client.Topic("topic1") msgIDs, err := topic.Publish(ctx, &pubsub.Message{ Data: []byte("hello world"), }) if err != nil { log.Fatal(err) } // Create an iterator to pull messages via subscription1. it, err := client.Subscription("subscription1").Pull(ctx) if err != nil { log.Println(err) } defer it.Stop() // Consume N messages from the iterator. for i := 0; i < N; i++ { msg, err := it.Next() if err == iterator.Done { break } if err != nil { log.Fatalf("Failed to retrieve message: %v", err) } fmt.Printf("Message %d: %s\n", i, msg.Data) msg.Done(true) // Acknowledge that we've consumed the message. }
First create a bigquery.Client
to use throughout your application:
c, err := bigquery.NewClient(ctx, "my-project-ID") if err != nil { // TODO: Handle error. }
Then use that client to interact with the API:
// Construct a query. q := c.Query(` SELECT year, SUM(number) FROM [bigquery-public-data:usa_names.usa_1910_2013] WHERE name = "William" GROUP BY year ORDER BY year `) // Execute the query. it, err := q.Read(ctx) if err != nil { // TODO: Handle error. } // Iterate through the results. for it.Next(ctx) { // Retrieve the current row into a list of values. var values bigquery.ValueList if err := it.Get(&values); err != nil { // TODO: Handle error. } fmt.Println(values) } if it.Err() != nil { // TODO: Handle it.Err() }
First create a logging.Client
to use throughout your application:
ctx := context.Background() client, err := logging.NewClient(ctx, "my-project") if err != nil { // TODO: Handle error. }
Usually, you'll want to add log entries to a buffer to be periodically flushed (automatically and asynchronously) to the Stackdriver Logging service.
logger := client.Logger("my-log")
logger.Log(logging.Entry{Payload: "something happened!"})
Close your client before your program exits, to flush any buffered log entries.
err = client.Close() if err != nil { // TODO: Handle error. }
Contributions are welcome. Please, see the CONTRIBUTING document for details. We‘re using Gerrit for our code reviews. Please don’t open pull requests against this repo, new pull requests will be automatically closed.
Please note that this project is released with a Contributor Code of Conduct. By participating in this project you agree to abide by its terms. See Contributor Code of Conduct for more information.