super

CLI for the Serverless Supercomputer

24
10
Ruby

Installation | Blogs | Tutorials

Super: A CLI for the Serverless Supercomputer

Super offers a zero-config and zero-code entry to your Cloud. It
does so by running normal UNIX command lines against Cloud data,
using Cloud compute. Super takes care of hooking these complex and
disparate resources together under one command: super run.

🚀 Take me to the Installation Instructions

Blogs

A Super Way to Copy

For example, Super can copy a set of files from one place in the
Cloud to another.

super run -- cp /s3/ibm/default/src/foo*.txt /s3/aws/dst

Behind the scenes, Super spawns Cloud Compute to mediate the
Cloud-to-Cloud data transfer. It uses
“glob” patterns to
determine how many concurrent jobs to run. For example, if foo*.txt
matches 5 files, Super spawns a set of concurrent jobs, grants each
job the least privilege to access its assigned files, and more!

Because Super intelligently parses your command line, it can
automatically inject progress trackers. Super tracks the progress of
any job
against your Cloud data.

Animated GIF of super copy

Super leverges any Kubernetes cluster for Compute and any S3
provider
for Data. If you wish to target a very large cluster, Super
integrates with IBM Cloud Code
Engine
. It also can hook your
Compute jobs up with IBM Cloud Object
Storage
. The super up command gives you an easy way to
leverage both.

There is no need to code to the Cloud API of the week to make any of
this happen.

What Other Kinds of Pipelines Can Super Run?

Click on an image for more detail on that use case.






Installation

macOSOthers
brew tap IBM/super https://github.com/IBM/super
brew install super
super
Coming soon

You should now see usage information for Super, including the main
sub-command: super run.

Getting Started: Using super run to submit pipelines to Cloud Compute

Out of the box, super run will use your current Kubernetes
context
as the target for Compute, and will have read-only access to
public S3 buckets.

If this works for you, then try super run -p5 -- echo hello. Above, we used a glob pattern to specify the
Cloud jobs we needed; here, since we are not pulling in Cloud data, we
instead use -p5 to specify that we want to execute the given command
line as five Cloud jobs.

Super auto-scales normal UNIX command lines

Using super browse to explore input Cloud Data

To browse for interesting CommonCrawl
input data, you may use super browse cc. Super pipelines can access
S3 data via a pseudo /s3 filepath; e.g. /s3/aws/commoncrawl is the
prefix for CommonCrawl data.

Using super target to configure your choice of Cloud Compute

By default, super run will target jobs against your currently
selected Kubernetes context. You may switch contexts using standard
kubectl commands. Strictly for convenience, Super offers super target to smooth the enumeration and selection of a context. In
particular, if you are using IBM Cloud Code Engine, the super target
command seamlessly integrates with CodeEngine projects.

Using super up to connect to your Cloud Provider

The super up command helps you with prerequisites

The super up command will attempt to
connect super run to your AWS credentials and to IBM Cloud. The
latter allows super run to scale to a large Kubernetes cluster with
hundreds of nodes, via IBM Cloud Code
Engine
; super up can also
connect you to your IBM Cloud Object
Storage
instances.

Tutorials