Skip to content

Commit 6539a3e

Browse files
committed
Use new tip shortcode everywhere
1 parent 7ba51dc commit 6539a3e

24 files changed

+97
-98
lines changed

docs/commands/super-db.md

+14-14
Original file line numberDiff line numberDiff line change
@@ -15,7 +15,7 @@ title: super db
1515
1616
<p id="status"></p>
1717

18-
:::tip Status
18+
{{< tip "Status" >}}
1919
While [`super`](super.md) and its accompanying [formats](../formats/_index.md)
2020
are production quality, the SuperDB data lake is still fairly early in development
2121
and alpha quality.
@@ -25,7 +25,7 @@ is deployed to manage the lake's data layout via the
2525
[lake API](../lake/api.md).
2626

2727
Enhanced scalability with self-tuning configuration is under development.
28-
:::
28+
{{< /tip >}}
2929

3030
## The Lake Model
3131

@@ -153,7 +153,7 @@ running any `super db` lake command all pointing at the same storage endpoint
153153
and the lake's data footprint will always remain consistent as the endpoints
154154
all adhere to the consistency semantics of the lake.
155155

156-
:::tip caveat
156+
{{< tip "Caveat" >}}
157157
Data consistency is not fully implemented yet for
158158
the S3 endpoint so only single-node access to S3 is available right now,
159159
though support for multi-node access is forthcoming.
@@ -164,7 +164,7 @@ access to a local file system has been thoroughly tested and should be
164164
deemed reliable, i.e., you can run a direct-access instance of `super db` alongside
165165
a server instance of `super db` on the same file system and data consistency will
166166
be maintained.
167-
:::
167+
{{< /tip >}}
168168

169169
### Locating the Lake
170170

@@ -206,11 +206,11 @@ Each commit object is assigned a global ID.
206206
Similar to Git, commit objects are arranged into a tree and
207207
represent the entire commit history of the lake.
208208

209-
:::tip note
209+
{{< tip "Note" >}}
210210
Technically speaking, Git can merge from multiple parents and thus
211211
Git commits form a directed acyclic graph instead of a tree;
212212
SuperDB does not currently support multiple parents in the commit object history.
213-
:::
213+
{{< /tip >}}
214214

215215
A branch is simply a named pointer to a commit object in the lake
216216
and like a pool, a branch name can be any valid UTF-8 string.
@@ -272,10 +272,10 @@ key. For example, on a pool with pool key `ts`, the query `ts == 100`
272272
will be optimized to scan only the data objects where the value `100` could be
273273
present.
274274

275-
:::tip note
275+
{{< tip "Note" >}}
276276
The pool key will also serve as the primary key for the forthcoming
277277
CRUD semantics.
278-
:::
278+
{{< /tip >}}
279279

280280
A pool also has a configured sort order, either ascending or descending
281281
and data is organized in the pool in accordance with this order.
@@ -325,9 +325,9 @@ using that pool's "branches log" in a similar fashion, then its corresponding
325325
commit object can be used to construct the data of that branch at that
326326
past point in time.
327327

328-
:::tip note
328+
{{< tip "Note" >}}
329329
Time travel using timestamps is a forthcoming feature.
330-
:::
330+
{{< /tip >}}
331331

332332
## `super db` Commands
333333

@@ -407,11 +407,11 @@ the [special value `this`](../language/pipeline-model.md#the-special-value-this)
407407

408408
A newly created pool is initialized with a branch called `main`.
409409

410-
:::tip note
410+
{{< tip "Note" >}}
411411
Lakes can be used without thinking about branches. When referencing a pool without
412412
a branch, the tooling presumes the "main" branch as the default, and everything
413413
can be done on main without having to think about branching.
414-
:::
414+
{{< /tip >}}
415415

416416
### Delete
417417
```
@@ -582,9 +582,9 @@ that is stored in the commit journal for reference. These values may
582582
be specified as options to the [`load`](#load) command, and are also available in the
583583
[lake API](../lake/api.md) for automation.
584584

585-
:::tip note
585+
{{< tip "Note" >}}
586586
The branchlog meta-query source is not yet implemented.
587-
:::
587+
{{< /tip >}}
588588

589589
### Ls
590590
```

docs/commands/super.md

+2-2
Original file line numberDiff line numberDiff line change
@@ -187,13 +187,13 @@ not desirable because (1) the Super JSON parser is not particularly performant a
187187
(2) all JSON numbers are floating point but the Super JSON parser will parse as
188188
JSON any number that appears without a decimal point as an integer type.
189189

190-
:::tip note
190+
{{< tip "Note" >}}
191191
The reason `super` is not particularly performant for Super JSON is that the [Super Binary](../formats/bsup.md) or
192192
[Super Columnar](../formats/csup.md) formats are semantically equivalent to Super JSON but much more efficient and
193193
the design intent is that these efficient binary formats should be used in
194194
use cases where performance matters. Super JSON is typically used only when
195195
data needs to be human-readable in interactive settings or in automated tests.
196-
:::
196+
{{< /tip >}}
197197

198198
To this end, `super` uses a heuristic to select between Super JSON and plain JSON when the
199199
`-i` option is not specified. Specifically, plain JSON is selected when the first values

docs/formats/bsup.md

+4-4
Original file line numberDiff line numberDiff line change
@@ -130,7 +130,7 @@ size decompression buffers in advance of decoding.
130130
Values for the `format` byte are defined in the
131131
[Super Binary compression format specification](./compression.md).
132132

133-
:::tip note
133+
{{< tip "Note" >}}
134134
This arrangement of frames separating types and values allows
135135
for efficient scanning and parallelization. In general, values depend
136136
on type definitions but as long as all of the types are known when
@@ -143,7 +143,7 @@ heuristics, e.g., knowing a filtering predicate can't be true based on a
143143
quick scan of the data perhaps using the Boyer-Moore algorithm to determine
144144
that a comparison with a string constant would not work for any
145145
value in the buffer.
146-
:::
146+
{{< /tip >}}
147147

148148
Whether the payload was originally uncompressed or was decompressed, it is
149149
then interpreted according to the `T` bits of the frame code as a
@@ -211,12 +211,12 @@ is further encoded as a "counted string", which is the `uvarint` encoding
211211
of the length of the string followed by that many bytes of UTF-8 encoded
212212
string data.
213213

214-
:::tip note
214+
{{< tip "Note" >}}
215215
As defined by [Super JSON](jsup.md), a field name can be any valid UTF-8 string much like JSON
216216
objects can be indexed with arbitrary string keys (via index operator)
217217
even if the field names available to the dot operator are restricted
218218
by language syntax for identifiers.
219-
:::
219+
{{< /tip >}}
220220

221221
The type ID follows the field name and is encoded as a `uvarint`.
222222

docs/formats/csup.md

+18-18
Original file line numberDiff line numberDiff line change
@@ -64,12 +64,12 @@ then write the metadata into the reassembly section along with the trailer
6464
at the end. This allows a stream to be converted to a Super Columnar file
6565
in a single pass.
6666

67-
:::tip note
67+
{{< tip "Note" >}}
6868
That said, the layout is
6969
flexible enough that an implementation may optimize the data layout with
7070
additional passes or by writing the output to multiple files then
7171
merging them together (or even leaving the Super Columnar entity as separate files).
72-
:::
72+
{{< /tip >}}
7373

7474
### The Data Section
7575

@@ -85,17 +85,17 @@ There is no information in the data section for how segments relate
8585
to one another or how they are reconstructed into columns. They are just
8686
blobs of Super Binary data.
8787

88-
:::tip note
88+
{{< tip "Note" >}}
8989
Unlike Parquet, there is no explicit arrangement of the column chunks into
9090
row groups but rather they are allowed to grow at different rates so a
9191
high-volume column might be comprised of many segments while a low-volume
9292
column must just be one or several. This allows scans of low-volume record types
9393
(the "mice") to perform well amongst high-volume record types (the "elephants"),
9494
i.e., there are not a bunch of seeks with tiny reads of mice data interspersed
9595
throughout the elephants.
96-
:::
96+
{{< /tip >}}
9797

98-
:::tip TBD
98+
{{< tip "TBD" >}}
9999
The mice/elephants model creates an interesting and challenging layout
100100
problem. If you let the row indexes get too far apart (call this "skew"), then
101101
you have to buffer very large amounts of data to keep the column data aligned.
@@ -109,15 +109,15 @@ if you use lots of buffering on ingest, you can write the mice in front of the
109109
elephants so the read path requires less buffering to align columns. Or you can
110110
do two passes where you store segments in separate files then merge them at close
111111
according to an optimization plan.
112-
:::
112+
{{< /tip >}}
113113

114114
### The Reassembly Section
115115

116116
The reassembly section provides the information needed to reconstruct
117117
column streams from segments, and in turn, to reconstruct the original values
118118
from column streams, i.e., to map columns back to composite values.
119119

120-
:::tip note
120+
{{< tip "Note" >}}
121121
Of course, the reassembly section also provides the ability to extract just subsets of columns
122122
to be read and searched efficiently without ever needing to reconstruct
123123
the original rows. How well this performs is up to any particular
@@ -127,7 +127,7 @@ Also, the reassembly section is in general vastly smaller than the data section
127127
so the goal here isn't to express information in cute and obscure compact forms
128128
but rather to represent data in an easy-to-digest, programmer-friendly form that
129129
leverages Super Binary.
130-
:::
130+
{{< /tip >}}
131131

132132
The reassembly section is a Super Binary stream. Unlike Parquet,
133133
which uses an externally described schema
@@ -147,9 +147,9 @@ A super type's integer position in this sequence defines its identifier
147147
encoded in the [super column](#the-super-column). This identifier is called
148148
the super ID.
149149

150-
:::tip note
150+
{{< tip "Note" >}}
151151
Change the first N values to type values instead of nulls?
152-
:::
152+
{{< /tip >}}
153153

154154
The next N+1 records contain reassembly information for each of the N super types
155155
where each record defines the column streams needed to reconstruct the original
@@ -171,11 +171,11 @@ type signature:
171171
In the rest of this document, we will refer to this type as `<segmap>` for
172172
shorthand and refer to the concept as a "segmap".
173173

174-
:::tip note
174+
{{< tip "Note" >}}
175175
We use the type name "segmap" to emphasize that this information represents
176176
a set of byte ranges where data is stored and must be read from *rather than*
177177
the data itself.
178-
:::
178+
{{< /tip >}}
179179

180180
#### The Super Column
181181

@@ -216,11 +216,11 @@ This simple top-down arrangement, along with the definition of the other
216216
column structures below, is all that is needed to reconstruct all of the
217217
original data.
218218

219-
:::tip note
219+
{{< tip "Note" >}}
220220
Each row reassembly record has its own layout of columnar
221221
values and there is no attempt made to store like-typed columns from different
222222
schemas in the same physical column.
223-
:::
223+
{{< /tip >}}
224224

225225
The notation `<any_column>` refers to any instance of the five column types:
226226
* [`<record_column>`](#record-column),
@@ -296,9 +296,9 @@ in the same column order implied by the union type, and
296296
* `tags` is a column of `int32` values where each subsequent value encodes
297297
the tag of the union type indicating which column the value falls within.
298298

299-
:::tip TBD
299+
{{< tip "TBD" >}}
300300
Change code to conform to columns array instead of record{c0,c1,...}
301-
:::
301+
{{< /tip >}}
302302

303303
The number of times each value of `tags` appears must equal the number of values
304304
in each respective column.
@@ -350,14 +350,14 @@ data in the file,
350350
it will typically fit comfortably in memory and it can be very fast to scan the
351351
entire reassembly structure for any purpose.
352352

353-
:::tip Example
353+
{{< tip "Example" >}}
354354
For a given query, a "scan planner" could traverse all the
355355
reassembly records to figure out which segments will be needed, then construct
356356
an intelligent plan for reading the needed segments and attempt to read them
357357
in mostly sequential order, which could serve as
358358
an optimizing intermediary between any underlying storage API and the
359359
Super Columnar decoding logic.
360-
:::
360+
{{< /tip >}}
361361

362362
To decode the "next" row, its schema index is read from the root reassembly
363363
column stream.

docs/install.md

+2-2
Original file line numberDiff line numberDiff line change
@@ -40,11 +40,11 @@ This installs the `super` binary in your `$GOPATH/bin`.
4040

4141
Once installed, run a [quick test](#quick-tests).
4242

43-
:::tip note
43+
{{< tip "Note" >}}
4444
If you don't have Go installed, download and install it from the
4545
[Go install page](https://golang.org/doc/install). Go 1.23 or later is
4646
required.
47-
:::
47+
{{< /tip >}}
4848

4949
## Quick Tests
5050

docs/integrations/amazon-s3.md

+2-2
Original file line numberDiff line numberDiff line change
@@ -16,11 +16,11 @@ You must specify an AWS region via one of the following:
1616
You can create `~/.aws/config` by installing the
1717
[AWS CLI](https://aws.amazon.com/cli/) and running `aws configure`.
1818

19-
:::tip Note
19+
{{< tip "Note" >}}
2020
If using S3-compatible storage that does not recognize the concept of regions,
2121
a region must still be specified, e.g., by providing a dummy value for
2222
`AWS_REGION`.
23-
:::
23+
{{< /tip >}}
2424

2525
## Credentials
2626

docs/integrations/fluentd.md

+5-4
Original file line numberDiff line numberDiff line change
@@ -81,13 +81,14 @@ The default settings when running `zed create` set the
8181
field and sort the stored data in descending order by that key. This
8282
configuration is ideal for Zeek log data.
8383

84-
:::tip Note
84+
{{< tip "Note" >}}
8585
The [Zui](https://zui.brimdata.io/) desktop application automatically starts a
8686
Zed lake service when it launches. Therefore if you are using Zui you can
8787
skip the first set of commands shown above. The pool can be created from Zui
8888
by clicking **+**, selecting **New Pool**, then entering `ts` for the
8989
[pool key](../commands/super-db.md#pool-key).
90-
:::
90+
{{< /tip >}}
91+
9192

9293
### Fluentd
9394

@@ -366,15 +367,15 @@ leverage, you can reduce the lake's storage footprint by periodically running
366367
storage that contain the granular commits that have already been rolled into
367368
larger objects by compaction.
368369

369-
:::tip Note
370+
{{< tip "Note" >}}
370371
As described in issue [super/4934](https://github.com/brimdata/super/issues/4934),
371372
even after running `zed vacuum`, some files related to commit history are
372373
currently still left behind below the lake storage path. The issue describes
373374
manual steps that can be taken to remove these files safely, if desired.
374375
However, if you find yourself needing to take these steps in your environment,
375376
please [contact us](#contact-us) as it will allow us to boost the priority
376377
of addressing the issue.
377-
:::
378+
{{< /tip >}}
378379

379380
## Ideas For Enhancement
380381

docs/integrations/zed-lake-auth/index.md

+4-4
Original file line numberDiff line numberDiff line change
@@ -30,10 +30,10 @@ and then clicking the **Create API** button.
3030

3131
2. Enter any **Name** and URL **Identifier** for the API, then click the
3232
**Create** button.
33-
:::tip
33+
{{< tip "Tip" >}}
3434
Note the value you enter for the **Identifier** as you'll
3535
need it later for the Zed lake service configuration.
36-
:::
36+
{{< /tip >}}
3737

3838
![api-name-identifier](api-name-identifier.png)
3939

@@ -50,11 +50,11 @@ need it later for the Zed lake service configuration.
5050

5151
1. Begin creating a new application by clicking **Applications** in the left
5252
navigation menu and then clicking the **Create Application** button.
53-
:::tip Note
53+
{{< tip "Note" >}}
5454
Neither the "Zed lake (Test Application)" that was created for us
5555
automatically when we created our API nor the Default App that came with the
5656
trial are used in this configuration.
57-
:::
57+
{{< /tip >}}
5858

5959
![create-application](create-application.png)
6060

docs/integrations/zeek/data-type-compatibility.md

+2-2
Original file line numberDiff line numberDiff line change
@@ -49,15 +49,15 @@ applicable to handling certain types.
4949
| [`vector`](https://docs.zeek.org/en/current/script-reference/types.html#type-vector) | [`array`](../../formats/zed.md#22-array | |
5050
| [`record`](https://docs.zeek.org/en/current/script-reference/types.html#type-record) | [`record`](../../formats/zed.md#21-record | See [`record` details](#record) |
5151

52-
:::tip Note
52+
{{< tip "Note" >}}
5353
The [Zeek data types](https://docs.zeek.org/en/current/script-reference/types.html)
5454
page describes the types in the context of the
5555
[Zeek scripting language](https://docs.zeek.org/en/master/scripting/index.html).
5656
The Zeek types available in scripting are a superset of the data types that
5757
may appear in Zeek log files. The encodings of the types also differ in some
5858
ways between the two contexts. However, we link to this reference because
5959
there is no authoritative specification of the Zeek TSV log format.
60-
:::
60+
{{< /tip >}}
6161

6262
## Example
6363

0 commit comments

Comments
 (0)