Skip to content

aloneguid/stowage

Folders and files

NameName
Last commit message
Last commit date

Latest commit

Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 

Repository files navigation

Stowage

Nuget Nuget (with prereleases) Nuget

logo

This documentation is for Stowage v2 which is a major redesign. Version 1 documentation can be found here.

Stowage is a bloat-free .NET cloud storage kit that supports at minimum THE major ☁ providers.

  • Independent πŸ†“. Provides an independent implementation of the ☁ storage APIs. Because you can't just have official corporate SDKs as a single source of truth.
  • Readable. Official SDKs like the ones for AWS, Google, or Azure are overengineered and unreadable. Some are autogenerated and look just bad and foreign to .NET ecosystem. Some won't even compile without some custom rituals.
  • Beautiful πŸ¦‹. Designed to fit into .NET ecosystem, not the other way around.
  • Rich πŸ’°. Provide maximum functionality. However, in addition to that, provide humanly possible way to easily extend it with new functionality, without waiting for new SDK releases.
  • Embeddable πŸ”±. Has zero external dependencies, relies only on built-in .NET API. Often official SDKs have a very deep dependency tree causing a large binary sizes and endless conflicts during runtime. This one is a single .NET .dll with no dependencies whatsoever.
  • Cross Cloud πŸŒ₯. Same API. Any cloud. Best decisions made for you. It's like iPhone vs Windows Phone.
  • Cross Tested ❎. It's not just cross cloud but also cross tested (I don't know how to call this). It tests that all cloud providers behave absolutely the same on various method calls. They should validate arguments the same, throw same exceptions in the same situations, and support the same set of functionality. Sounds simple, but it's rare to find in a library. And it important, otherwise what's the point of a generic API if you need to write a lot of if()s? (or pattern matching).

This library originally came out from being frustrated on working on my another library - Storage.Net. While it's OK, most of the time I had to deal with SDK incompatibilities, breaking changes, oddnesses, and slowness, whereas most of the time users needs something simple that just works.

Getting Started

Right, time to gear up. We'll do it step by step. First, you need to install the Nuget package.

Simplest case, using the local πŸ’½ and writing text "I'm a page!!!" to a file called "pagefile.sys" at the root of disk C::

using Stowage;

using(IFileStorage fs = Files.Of.LocalDisk("c:\\")) {
   await fs.WriteText("pagefile.sys", "I'm a page!!!!");
}

This is local disk, yeah? But what about cloud storage, like Azure Blob Storage? Piece of cake:

using Stowage;

using(IFileStorage fs = Files.Of.AzureBlobStorage("accountName", "accountKey", "containerName")) {
   var entries = await fs.Ls();
}

β™’ Streaming

Streaming is a first-class feature. This means the streaming is real with no workarounds or in-memory buffers, so you can upload/download files of virtually unlimited sizes. Most official SDKs do not support streaming at all - surprisingly even the cloud leader's .NET SDK doesn't. Each requires some sort of crippled down version of stream - either knowing length beforehand, or will buffer it all in memory. I don't. I stream like a stream.

Proper streaming support also means that you can transform streams as you write to them or read from them - something that is not available in the native SDKs. For instance gzipping, encryption, anything else.

Streaming is also truly compatible with synchronous and asynchronous API.

Details/Documentation

Whenever a method appears here, I assume it belongs to IFileStorage interface, unless specified.

Listing/Browsing

Use .Ls() (short for list) - very easy to remember! Everyone knows what ls does, right? Optionally allows to list entries recursively.

Reading

The core method for reading is Stream OpenRead(IOPath path) - this returns a stream from file path. Stream is the lowest level data structure. There are other helper methods that by default rely on this method, like ReadText etc. Just have a quick look:

IFileStorage fs = ...;
Stream target = ...;

// copy to another stream
using Stream s = await fs.OpenRead("/myfile.txt");

// synchronous copy:
s.CopyTo(target);

// or alternatively, asynchronous copy (preferred):
await s.CopyToAsync(target);

// if you just need text:
string content = await fs.ReadText("/myfile.txt");

Of course there are more overloaded methods you can take advantage of.

Writing

The main method Stream OpenWrite(IOPath path, ...) opens(/creates?) a file for writing. It returns a real writeable stream you can write to and close afterwards. It behaves like a stream and is a stream.

There are other overloads which support writing text etc.

Destroying 🧨

Rm(IOPath path) trashes files or folders (or both) with options to do it recursively!

Other

There are other useful utility methods:

  • bool Exists(IOPath path) that checks for file existence. It supposed to be really efficient, hence a separate method.
  • Ren renames files and folders.
  • and more are coming - check IFileStorage interface to be up to date.

Supported Storage Systems (Built-In)

Instantiation instructions are in the code documentation (IntelliSense?) - I prefer this to writing out here locally.

Below are some details worth mentioning.

Local disk

It's just what it says - local disk, which by default maps to an entire filesystem. In Windows, which has drive letters, it will map to an entire disk of where the application's current directory's drive is (i.e. if you are in c:/my/app it it will map to c:/).

IFileStorage storage = Files.Of.LocalDisk();
// or
IFileStorage storage = Storage.Of.ConnectionString("disk://")

Optionally, you can specify the root directory:

IFileStorage storage = Files.Of.LocalDisk("/a/folder");
// or
IFileStorage storage = Storage.Of.ConnectionString("disk://path=/a/folder")

do not forget there is path keyword in the connection string, it's so tempting to write "disk://a/folder" instead of "disk://path=/a/folder"!

AWS S3

In AWS, the path addressing style is the following:

/bucket/path/object

Ls on the root folder returns list of buckets in the AWS account, whether you do have access to them or not.

Authentication

Key/Secret

The most usual way to authenticate with S3 is to use the following method:

IFileStorage storage = Files.Of.AmazonS3(key, secret, region);
// or
IFileStorage storage = Storage.Of.ConnectionString("s3://keyId=<key>;key=<secret>;region=<region>")

These are what Amazon calls "long-term" credentials. If you are using STS, the same method overload allows you to pass sessionToken.

IFileStorage storage = Files.Of.AmazonS3(key, secret, region, sessionToken);
//or
IFileStorage storage = Storage.Of.ConnectionString("s3://keyId=<key>;key=<secret>;region=<region>;sessionToken=<session token>")
CLI Profile

Another way to authenticate is using CLI profile. This is useful when you machine is already authenticated using aws cli, awsume or similar tools that write credentials and configuration to ~/.aws/credentials and ~/.aws/config.

You only need to pass the profile name (and only if it's not a default one):

// ---- using default profile ----

IFileStorage storage = Files.Of.AmazonS3FromCliProfile();
//or
IFileStorage storage = Storage.Of.ConnectionString("s3://");

// ----- using specific profile, like "myprofile" ----
IFileStorage storage = Files.Of.AmazonS3FromCliProfile("myprofile");
//or
IFileStorage storage = Storage.Of.ConnectionString("s3://profile=myprofile");

This method has other default parameters, such as regionName which can be specified or overridden if not found in CLI configuration, i.e. Files.Of.AmazonS3FromCliProfile(); has optional region parameters, and connection string has optional region= keyword.

Minio

Minio is essentially using the standard S3 protocol, but addressing style is slightly different. There is a helper extension that somewhat simplifies Minio authentication:

IFileStorage storage = Files.Of.Minio(endpoint, key, secret);

Azure Blob Storage

In Azure Blob Storage, path addressing style is the following:

/container/path/object

Note that there is no storage account in the path, mostly because Shared Key authentication is storage account scoped, not tenant scoped.

Ls on the root folder returns list of containers in the storage account.

Authentication

Shared Key

Azure provider supports authentication with Shared Key:

IFileStorage storage = Files.Of.AzureBlobStorage(accountName, sharedKey);
// or
IFileStorage storage = Storage.Of.ConnectionString("az://account=<account name>;key=<shared key>")
Entra Id service principals
IFileStorage storage = Files.Of.AzureBlobStorage(
    accountName,
    new ClientSecretCredential(tenantId, clientId, clientSecret));

Managed identities are not yet supported due to low demand, but watch this space.

Emulator

Azure emulator is supported, just use AzureBlobStorageWithLocalEmulator() method to connect to it.

Exotic providers

Local disk cache

This storage essentially wraps around another storage to provide content caching capabilities. Example:

IFileStorage storage = Files.Of.AzureBlobStorage(accountName, sharedKey);
IFileStorage cachedStorage = Files.Of.LocalDiskCacheStorage(storage);

When using cachedStorage, all the operations are forwarded to storage as is, except for OpenRead which downloads content locally and opens a stream to the local file.

πŸ“ˆ Extending

There are many ways to extend functionality:

  1. Documentation. You might think it's not extending anything, however if user is not aware for some functionality it doesn't exist. Documenting it is making it available, hence extending. You must be slightly mad to follow my style of writing though.
  2. New functionality. Adding utility methods like copying files inside or between accounts, automatic JSON serialisation etc. is always good. Look IFileStorage interface and PolyfilledFileStorage. In most cases these two files are enough to add pure business logic. Not counting unit tests. Which you must write. Otherwise it's easier to do the whole thing by myself. Which is what will happen according to my experience.
  3. Native optimisations. Some functionality is generic, and some depends on a specific cloud provider. For instance, one can copy a file by downloading it locally, and uploading with a new name. Or utilise a native REST call that accepts source and target file name, if it exists. Involves digging deeper into specific provider's API.

When contributing a new provider, it's way more preferrable to embed it's code in the library, provided that:

  • there are no extra nuget dependencies.
  • it's cross-platform.

I'm a strong advocate of simplicity and not going to repeat the mistake of turning this into a nuget tree dependency hell!

❔ Who?

Raise a PR to appear here.

Related Projects

  • RCLONE - cross-platform open-source cloud sync tool.
  • Storage.Net - the roots of this project.

πŸ’° Contributing

You are welcome to contribute in any form, however I wouldn't bother, especially financially. Don't bother buying me a β˜•, I can do it myself real cheap. During my years of OSS development everyone I know (including myself) have only lost money. Why I'm still doing this? Probably because it's just cool and I'm enjoying it.