You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: docs/dyn/firestore_v1.projects.databases.backupSchedules.html
+6-6Lines changed: 6 additions & 6 deletions
Original file line number
Diff line number
Diff line change
@@ -112,7 +112,7 @@ <h3>Method Details</h3>
112
112
"dailyRecurrence": { # Represent a recurring schedule that runs at a specific time every day. The time zone is UTC. # For a schedule that runs daily at a specified time.
113
113
},
114
114
"name": "A String", # Output only. The unique backup schedule identifier across all locations and databases for the given project. This will be auto-assigned. Format is `projects/{project}/databases/{database}/backupSchedules/{backup_schedule}`
115
-
"retention": "A String", # At what relative time in the future, compared to the creation time of the backup should the backup be deleted, i.e. keep backups for 7 days.
115
+
"retention": "A String", # At what relative time in the future, compared to its creation time, the backup should be deleted, e.g. keep backups for 7 days.
116
116
"updateTime": "A String", # Output only. The timestamp at which this backup schedule was most recently updated. When a backup schedule is first created, this is the same as create_time.
117
117
"weeklyRecurrence": { # Represents a recurring schedule that runs on a specified day of the week. The time zone is UTC. # For a schedule that runs weekly on a specific day and time.
118
118
"day": "A String", # The day of week to run. DAY_OF_WEEK_UNSPECIFIED is not allowed.
@@ -132,7 +132,7 @@ <h3>Method Details</h3>
132
132
"dailyRecurrence": { # Represent a recurring schedule that runs at a specific time every day. The time zone is UTC. # For a schedule that runs daily at a specified time.
133
133
},
134
134
"name": "A String", # Output only. The unique backup schedule identifier across all locations and databases for the given project. This will be auto-assigned. Format is `projects/{project}/databases/{database}/backupSchedules/{backup_schedule}`
135
-
"retention": "A String", # At what relative time in the future, compared to the creation time of the backup should the backup be deleted, i.e. keep backups for 7 days.
135
+
"retention": "A String", # At what relative time in the future, compared to its creation time, the backup should be deleted, e.g. keep backups for 7 days.
136
136
"updateTime": "A String", # Output only. The timestamp at which this backup schedule was most recently updated. When a backup schedule is first created, this is the same as create_time.
137
137
"weeklyRecurrence": { # Represents a recurring schedule that runs on a specified day of the week. The time zone is UTC. # For a schedule that runs weekly on a specific day and time.
138
138
"day": "A String", # The day of week to run. DAY_OF_WEEK_UNSPECIFIED is not allowed.
@@ -177,7 +177,7 @@ <h3>Method Details</h3>
177
177
"dailyRecurrence": { # Represent a recurring schedule that runs at a specific time every day. The time zone is UTC. # For a schedule that runs daily at a specified time.
178
178
},
179
179
"name": "A String", # Output only. The unique backup schedule identifier across all locations and databases for the given project. This will be auto-assigned. Format is `projects/{project}/databases/{database}/backupSchedules/{backup_schedule}`
180
-
"retention": "A String", # At what relative time in the future, compared to the creation time of the backup should the backup be deleted, i.e. keep backups for 7 days.
180
+
"retention": "A String", # At what relative time in the future, compared to its creation time, the backup should be deleted, e.g. keep backups for 7 days.
181
181
"updateTime": "A String", # Output only. The timestamp at which this backup schedule was most recently updated. When a backup schedule is first created, this is the same as create_time.
182
182
"weeklyRecurrence": { # Represents a recurring schedule that runs on a specified day of the week. The time zone is UTC. # For a schedule that runs weekly on a specific day and time.
183
183
"day": "A String", # The day of week to run. DAY_OF_WEEK_UNSPECIFIED is not allowed.
@@ -206,7 +206,7 @@ <h3>Method Details</h3>
206
206
"dailyRecurrence": { # Represent a recurring schedule that runs at a specific time every day. The time zone is UTC. # For a schedule that runs daily at a specified time.
207
207
},
208
208
"name": "A String", # Output only. The unique backup schedule identifier across all locations and databases for the given project. This will be auto-assigned. Format is `projects/{project}/databases/{database}/backupSchedules/{backup_schedule}`
209
-
"retention": "A String", # At what relative time in the future, compared to the creation time of the backup should the backup be deleted, i.e. keep backups for 7 days.
209
+
"retention": "A String", # At what relative time in the future, compared to its creation time, the backup should be deleted, e.g. keep backups for 7 days.
210
210
"updateTime": "A String", # Output only. The timestamp at which this backup schedule was most recently updated. When a backup schedule is first created, this is the same as create_time.
211
211
"weeklyRecurrence": { # Represents a recurring schedule that runs on a specified day of the week. The time zone is UTC. # For a schedule that runs weekly on a specific day and time.
212
212
"day": "A String", # The day of week to run. DAY_OF_WEEK_UNSPECIFIED is not allowed.
@@ -230,7 +230,7 @@ <h3>Method Details</h3>
230
230
"dailyRecurrence": { # Represent a recurring schedule that runs at a specific time every day. The time zone is UTC. # For a schedule that runs daily at a specified time.
231
231
},
232
232
"name": "A String", # Output only. The unique backup schedule identifier across all locations and databases for the given project. This will be auto-assigned. Format is `projects/{project}/databases/{database}/backupSchedules/{backup_schedule}`
233
-
"retention": "A String", # At what relative time in the future, compared to the creation time of the backup should the backup be deleted, i.e. keep backups for 7 days.
233
+
"retention": "A String", # At what relative time in the future, compared to its creation time, the backup should be deleted, e.g. keep backups for 7 days.
234
234
"updateTime": "A String", # Output only. The timestamp at which this backup schedule was most recently updated. When a backup schedule is first created, this is the same as create_time.
235
235
"weeklyRecurrence": { # Represents a recurring schedule that runs on a specified day of the week. The time zone is UTC. # For a schedule that runs weekly on a specific day and time.
236
236
"day": "A String", # The day of week to run. DAY_OF_WEEK_UNSPECIFIED is not allowed.
@@ -251,7 +251,7 @@ <h3>Method Details</h3>
251
251
"dailyRecurrence": { # Represent a recurring schedule that runs at a specific time every day. The time zone is UTC. # For a schedule that runs daily at a specified time.
252
252
},
253
253
"name": "A String", # Output only. The unique backup schedule identifier across all locations and databases for the given project. This will be auto-assigned. Format is `projects/{project}/databases/{database}/backupSchedules/{backup_schedule}`
254
-
"retention": "A String", # At what relative time in the future, compared to the creation time of the backup should the backup be deleted, i.e. keep backups for 7 days.
254
+
"retention": "A String", # At what relative time in the future, compared to its creation time, the backup should be deleted, e.g. keep backups for 7 days.
255
255
"updateTime": "A String", # Output only. The timestamp at which this backup schedule was most recently updated. When a backup schedule is first created, this is the same as create_time.
256
256
"weeklyRecurrence": { # Represents a recurring schedule that runs on a specified day of the week. The time zone is UTC. # For a schedule that runs weekly on a specific day and time.
257
257
"day": "A String", # The day of week to run. DAY_OF_WEEK_UNSPECIFIED is not allowed.
Copy file name to clipboardExpand all lines: docs/dyn/firestore_v1.projects.databases.documents.html
+12-2Lines changed: 12 additions & 2 deletions
Original file line number
Diff line number
Diff line change
@@ -1398,9 +1398,9 @@ <h3>Method Details</h3>
1398
1398
],
1399
1399
},
1400
1400
"filter": { # A digest of all the documents that match a given target. # A filter to apply to the set of documents previously returned for the given target. Returned when documents may have been removed from the given target, but the exact documents are unknown.
1401
-
"count": 42, # The total count of documents that match target_id. If different from the count of documents in the client that match, the client must manually determine which documents no longer match the target. The client can use the `unchanged_names` bloom filter to assist with this determination.
1401
+
"count": 42, # The total count of documents that match target_id. If different from the count of documents in the client that match, the client must manually determine which documents no longer match the target. The client can use the `unchanged_names` bloom filter to assist with this determination by testing ALL the document names against the filter; if the document name is NOT in the filter, it means the document no longer matches the target.
1402
1402
"targetId": 42, # The target ID to which this filter applies.
1403
-
"unchangedNames": { # A bloom filter (https://en.wikipedia.org/wiki/Bloom_filter). The bloom filter hashes the entries with MD5 and treats the resulting 128-bit hash as 2 distinct 64-bit hash values, interpreted as unsigned integers using 2's complement encoding. These two hash values, named `h1` and `h2`, are then used to compute the `hash_count` hash values using the formula, starting at `i=0`: h(i) = h1 + (i * h2) These resulting values are then taken modulo the number of bits in the bloom filter to get the bits of the bloom filter to test for the given entry. # A bloom filter that contains the UTF-8 byte encodings of the resource names of the documents that match target_id, in the form `projects/{project_id}/databases/{database_id}/documents/{document_path}` that have NOT changed since the query results indicated by the resume token or timestamp given in `Target.resume_type`. This bloom filter may be omitted at the server's discretion, such as if it is deemed that the client will not make use of it or if it is too computationally expensive to calculate or transmit. Clients must gracefully handle this field being absent by falling back to the logic used before this field existed; that is, re-add the target without a resume token to figure out which documents in the client's cache are out of sync.
1403
+
"unchangedNames": { # A bloom filter (https://en.wikipedia.org/wiki/Bloom_filter). The bloom filter hashes the entries with MD5 and treats the resulting 128-bit hash as 2 distinct 64-bit hash values, interpreted as unsigned integers using 2's complement encoding. These two hash values, named `h1` and `h2`, are then used to compute the `hash_count` hash values using the formula, starting at `i=0`: h(i) = h1 + (i * h2) These resulting values are then taken modulo the number of bits in the bloom filter to get the bits of the bloom filter to test for the given entry. # A bloom filter that, despite its name, contains the UTF-8 byte encodings of the resource names of ALL the documents that match target_id, in the form `projects/{project_id}/databases/{database_id}/documents/{document_path}`. This bloom filter may be omitted at the server's discretion, such as if it is deemed that the client will not make use of it or if it is too computationally expensive to calculate or transmit. Clients must gracefully handle this field being absent by falling back to the logic used before this field existed; that is, re-add the target without a resume token to figure out which documents in the client's cache are out of sync.
1404
1404
"bits": { # A sequence of bits, encoded in a byte array. Each byte in the `bitmap` byte array stores 8 bits of the sequence. The only exception is the last byte, which may store 8 _or fewer_ bits. The `padding` defines the number of bits of the last byte to be ignored as "padding". The values of these "padding" bits are unspecified and must be ignored. To retrieve the first bit, bit 0, calculate: `(bitmap[0] & 0x01) != 0`. To retrieve the second bit, bit 1, calculate: `(bitmap[0] & 0x02) != 0`. To retrieve the third bit, bit 2, calculate: `(bitmap[0] & 0x04) != 0`. To retrieve the fourth bit, bit 3, calculate: `(bitmap[0] & 0x08) != 0`. To retrieve bit n, calculate: `(bitmap[n / 8] & (0x01 << (n % 8))) != 0`. The "size" of a `BitSequence` (the number of bits it contains) is calculated by this formula: `(bitmap.length * 8) - padding`. # The bloom filter data.
1405
1405
"bitmap": "A String", # The bytes that encode the bit sequence. May have a length of zero.
1406
1406
"padding": 42, # The number of bits of the last byte in `bitmap` to ignore as "padding". If the length of `bitmap` is zero, then this value must be `0`. Otherwise, this value must be between 0 and 7, inclusive.
@@ -1763,9 +1763,19 @@ <h3>Method Details</h3>
1763
1763
"aggregations": [ # Optional. Series of aggregations to apply over the results of the `structured_query`. Requires: * A minimum of one and maximum of five aggregations per query.
1764
1764
{ # Defines an aggregation that produces a single result.
1765
1765
"alias": "A String", # Optional. Optional name of the field to store the result of the aggregation into. If not provided, Firestore will pick a default name following the format `field_`. For example: ``` AGGREGATE COUNT_UP_TO(1) AS count_up_to_1, COUNT_UP_TO(2), COUNT_UP_TO(3) AS count_up_to_3, COUNT(*) OVER ( ... ); ``` becomes: ``` AGGREGATE COUNT_UP_TO(1) AS count_up_to_1, COUNT_UP_TO(2) AS field_1, COUNT_UP_TO(3) AS count_up_to_3, COUNT(*) AS field_2 OVER ( ... ); ``` Requires: * Must be unique across all aggregation aliases. * Conform to document field name limitations.
1766
+
"avg": { # Average of the values of the requested field. * Only numeric values will be aggregated. All non-numeric values including `NULL` are skipped. * If the aggregated values contain `NaN`, returns `NaN`. * If the aggregated value set is empty, returns `NULL`. * Always returns the result as a double. # Average aggregator.
1767
+
"field": { # A reference to a field in a document, ex: `stats.operations`. # The field to aggregate on.
1768
+
"fieldPath": "A String", # The relative path of the document being referenced. Requires: * Conform to document field name limitations.
1769
+
},
1770
+
},
1766
1771
"count": { # Count of documents that match the query. The `COUNT(*)` aggregation function operates on the entire document so it does not require a field reference. # Count aggregator.
1767
1772
"upTo": "A String", # Optional. Optional constraint on the maximum number of documents to count. This provides a way to set an upper bound on the number of documents to scan, limiting latency, and cost. Unspecified is interpreted as no bound. High-Level Example: ``` AGGREGATE COUNT_UP_TO(1000) OVER ( SELECT * FROM k ); ``` Requires: * Must be greater than zero when present.
1768
1773
},
1774
+
"sum": { # Sum of the values of the requested field. * Only numeric values will be aggregated. All non-numeric values including `NULL` are skipped. * If the aggregated values contain `NaN`, returns `NaN`. * If the aggregated value set is empty, returns 0. * Returns a 64-bit integer if the sum result is an integer value and does not overflow. Otherwise, the result is returned as a double. Note that even if all the aggregated values are integers, the result is returned as a double if it cannot fit within a 64-bit signed integer. When this occurs, the returned value will lose precision. * When underflow occurs, floating-point aggregation is non-deterministic. This means that running the same query repeatedly without any changes to the underlying values could produce slightly different results each time. In those cases, values should be stored as integers over floating-point numbers. # Sum aggregator.
1775
+
"field": { # A reference to a field in a document, ex: `stats.operations`. # The field to aggregate on.
1776
+
"fieldPath": "A String", # The relative path of the document being referenced. Requires: * Conform to document field name limitations.
1777
+
},
1778
+
},
1769
1779
},
1770
1780
],
1771
1781
"structuredQuery": { # A Firestore query. # Nested structured query.
0 commit comments