Skip to content

fix(deps): update module cloud.google.com/go/compute/metadata to v0.5.0 - autoclosed

Soos requested to merge renovate/cloud.google.com-go-compute-metadata-0.x into main

This MR contains the following updates:

Package Type Update Change
cloud.google.com/go/compute/metadata require minor v0.2.3 -> v0.5.0

Release Notes

googleapis/google-cloud-go (cloud.google.com/go/compute/metadata)

v0.5.0

Compare Source

  • bigquery:
    • The SQL types DATE, TIME and DATETIME are now supported. They correspond to the Date, Time and DateTime types in the new cloud.google.com/go/civil package.
    • Support for query parameters.
    • Support deleting a dataset.
    • Values from INTEGER columns will now be returned as int64, not int. This will avoid errors arising from large values on 32-bit systems.
  • datastore:
    • Nested Go structs encoded as Entity values, instead of a flattened list of the embedded struct's fields. This means that you may now have twice-nested slices, eg.
      type State struct {
        Cities  []struct{
          Populations []int
        }
      }
      See the announcement for more details.
    • Contexts no longer hold namespaces; instead you must set a key's namespace explicitly. Also, key functions have been changed and renamed.
    • The WithNamespace function has been removed. To specify a namespace in a Query, use the Query.Namespace method:
      q := datastore.NewQuery("Kind").Namespace("ns")
    • All the fields of Key are exported. That means you can construct any Key with a struct literal:
      k := &Key{Kind: "Kind",  ID: 37, Namespace: "ns"}
    • As a result of the above, the Key methods Kind, ID, d.Name, Parent, SetParent and Namespace have been removed.
    • NewIncompleteKey has been removed, replaced by IncompleteKey. Replace
      NewIncompleteKey(ctx, kind, parent)
      with
      IncompleteKey(kind, parent)
      and if you do use namespaces, make sure you set the namespace on the returned key.
    • NewKey has been removed, replaced by NameKey and IDKey. Replace
      NewKey(ctx, kind, name, 0, parent)
      NewKey(ctx, kind, "", id, parent)
      with
      NameKey(kind, name, parent)
      IDKey(kind, id, parent)
      and if you do use namespaces, make sure you set the namespace on the returned key.
    • The Done variable has been removed. Replace datastore.Done with iterator.Done, from the package google.golang.org/api/iterator.
    • The Client.Close method will have a return type of error. It will return the result of closing the underlying gRPC connection.
    • See the announcement for more details.

v0.4.0

Compare Source

  • bigquery: -NewGCSReference is now a function, not a method on Client.

    • Table.LoaderFrom now accepts a ReaderSource, enabling loading data into a table from a file or any io.Reader.
    • Client.Table and Client.OpenTable have been removed. Replace

      client.OpenTable("project", "dataset", "table")

      with

      client.DatasetInProject("project", "dataset").Table("table")
    • Client.CreateTable has been removed. Replace

      client.CreateTable(ctx, "project", "dataset", "table")

      with

      client.DatasetInProject("project", "dataset").Table("table").Create(ctx)
    • Dataset.ListTables have been replaced with Dataset.Tables. Replace

      tables, err := ds.ListTables(ctx)

      with

      it := ds.Tables(ctx)
      for {
          table, err := it.Next()
          if err == iterator.Done {
              break
          }
          if err != nil {
              // TODO: Handle error.
          }
          // TODO: use table.
      }
    • Client.Read has been replaced with Job.Read, Table.Read and Query.Read. Replace

      it, err := client.Read(ctx, job)

      with

      it, err := job.Read(ctx)

      and similarly for reading from tables or queries.

    • The iterator returned from the Read methods is now named RowIterator. Its behavior is closer to the other iterators in these libraries. It no longer supports the Schema method; see the next item. Replace

      for it.Next(ctx) {
          var vals ValueList
          if err := it.Get(&vals); err != nil {
              // TODO: Handle error.
          }
          // TODO: use vals.
      }
      if err := it.Err(); err != nil {
          // TODO: Handle error.
      }

      with for { var vals ValueList err := it.Next(&vals) if err == iterator.Done { break } if err != nil { // TODO: Handle error. } // TODO: use vals. } Instead of the RecordsPerRequest(n) option, write

      it.PageInfo().MaxSize = n

      Instead of the StartIndex(i) option, write

      it.StartIndex = i
    • ValueLoader.Load now takes a Schema in addition to a slice of Values. Replace

      func (vl *myValueLoader) Load(v []bigquery.Value)

      with

      func (vl *myValueLoader) Load(v []bigquery.Value, s bigquery.Schema)
    • Table.Patch is replace by Table.Update. Replace

      p := table.Patch()
      p.Description("new description")
      metadata, err := p.Apply(ctx)

      with

      metadata, err := table.Update(ctx, bigquery.TableMetadataToUpdate{
          Description: "new description",
      })
    • Client.Copy is replaced by separate methods for each of its four functions. All options have been replaced by struct fields.

      • To load data from Google Cloud Storage into a table, use Table.LoaderFrom.

        Replace

        client.Copy(ctx, table, gcsRef)

        with

        table.LoaderFrom(gcsRef).Run(ctx)

        Instead of passing options to Copy, set fields on the Loader:

        loader := table.LoaderFrom(gcsRef)
        loader.WriteDisposition = bigquery.WriteTruncate
      • To extract data from a table into Google Cloud Storage, use Table.ExtractorTo. Set fields on the returned Extractor instead of passing options.

        Replace

        client.Copy(ctx, gcsRef, table)

        with

        table.ExtractorTo(gcsRef).Run(ctx)
      • To copy data into a table from one or more other tables, use Table.CopierFrom. Set fields on the returned Copier instead of passing options.

        Replace

        client.Copy(ctx, dstTable, srcTable)

        with

        dst.Table.CopierFrom(srcTable).Run(ctx)
      • To start a query job, create a Query and call its Run method. Set fields on the query instead of passing options.

        Replace

        client.Copy(ctx, table, query)

        with

        query.Run(ctx)
    • Table.NewUploader has been renamed to Table.Uploader. Instead of options, configure an Uploader by setting its fields. Replace

      u := table.NewUploader(bigquery.UploadIgnoreUnknownValues())

      with

      u := table.NewUploader(bigquery.UploadIgnoreUnknownValues())
      u.IgnoreUnknownValues = true
  • pubsub: remove pubsub.Done. Use iterator.Done instead, where iterator is the package google.golang.org/api/iterator.

v0.3.0

  • storage:

    • AdminClient replaced by methods on Client. Replace

      adminClient.CreateBucket(ctx, bucketName, attrs)

      with

      client.Bucket(bucketName).Create(ctx, projectID, attrs)
    • BucketHandle.List replaced by BucketHandle.Objects. Replace

      for query != nil {
          objs, err := bucket.List(d.ctx, query)
          if err != nil { ... }
          query = objs.Next
          for _, obj := range objs.Results {
              fmt.Println(obj)
          }
      }

      with

      iter := bucket.Objects(d.ctx, query)
      for {
          obj, err := iter.Next()
          if err == iterator.Done {
              break
          }
          if err != nil { ... }
          fmt.Println(obj)
      }

      (The iterator package is at google.golang.org/api/iterator.)

      Replace Query.Cursor with ObjectIterator.PageInfo().Token.

      Replace Query.MaxResults with ObjectIterator.PageInfo().MaxSize.

    • ObjectHandle.CopyTo replaced by ObjectHandle.CopierFrom. Replace

      attrs, err := src.CopyTo(ctx, dst, nil)

      with

      attrs, err := dst.CopierFrom(src).Run(ctx)

      Replace

      attrs, err := src.CopyTo(ctx, dst, &storage.ObjectAttrs{ContextType: "text/html"})

      with

      c := dst.CopierFrom(src)
      c.ContextType = "text/html"
      attrs, err := c.Run(ctx)
    • ObjectHandle.ComposeFrom replaced by ObjectHandle.ComposerFrom. Replace

      attrs, err := dst.ComposeFrom(ctx, []*storage.ObjectHandle{src1, src2}, nil)

      with

      attrs, err := dst.ComposerFrom(src1, src2).Run(ctx)
    • ObjectHandle.Update's ObjectAttrs argument replaced by ObjectAttrsToUpdate. Replace

      attrs, err := obj.Update(ctx, &storage.ObjectAttrs{ContextType: "text/html"})

      with

      attrs, err := obj.Update(ctx, storage.ObjectAttrsToUpdate{ContextType: "text/html"})
    • ObjectHandle.WithConditions replaced by ObjectHandle.If. Replace

      obj.WithConditions(storage.Generation(gen), storage.IfMetaGenerationMatch(mgen))

      with

      obj.Generation(gen).If(storage.Conditions{MetagenerationMatch: mgen})

      Replace

      obj.WithConditions(storage.IfGenerationMatch(0))

      with

      obj.If(storage.Conditions{DoesNotExist: true})
    • storage.Done replaced by iterator.Done (from package google.golang.org/api/iterator).

  • Package preview/logging deleted. Use logging instead.


Configuration

📅 Schedule: Branch creation - "every weekend" (UTC), Automerge - At any time (no schedule defined).

🚦 Automerge: Disabled by config. Please merge this manually once you are satisfied.

Rebasing: Whenever MR becomes conflicted, or you tick the rebase/retry checkbox.

🔕 Ignore: Close this MR and you won't be reminded about this update again.


  • If you want to rebase/retry this MR, check this box

This MR has been generated by Renovate Bot.

Edited by Soos

Merge request reports