Your own private npm registry and backup server. Designed to be easy to set up and maintain, performant, and stable.
This project allows you to have your own npm registry. This server works with the necessary
npm commands just like the npmjs.org registry. You can use it to not worry about npm going down or to store your private packages. It performs much faster than npmjs.org and can even be matched with a CDN like Cloudfront to be fast globally.
Rather than trying to copy all the data in npm, this acts more like a proxy. While npm is up, it will cache package data locally or in S3. If npm goes down, it will deliver whatever is available in the cache. This means it won't be a fully comprehensive backup of npm, but you will be able to access anything you accessed before. This makes it easy to set up since you don't need to mirror the entire registry. Any packages previously accessed will be available.
The inspiration for this project comes from sinopia. This came out of a need for better cache, CDN, and general performance as well as stability of being able to run multiple instances without depending on a local filesystem.
This is also a 12 Factor app to make it easy to host on a PaaS like Heroku or in a custom Ansible/Chef/Puppet cluster.
The easiest way to set this up is with the Heroku button (you must use S3 with Heroku):
Alternatively, you can set it up from npm:
$ npm install -g npm-register $ npm-register
Either way, your registry is now setup and you should be able to test it by updating the packages with it:
$ npm update --registry http://urltomyregistry
See below for how to enable authorization and
Use S3 for storage by setting
NPM_REGISTER_STORAGE=s3. Then set
AWS_S3_BUCKET to the proper values.
Using the local filesystem is the default. You can explicitly set the storage with
NPM_REGISTER_STORAGE=fs. Select the location for the files to be stored with
NPM_REGISTER_FS_DIRECTORY=/var/npm-register. Defaults to
Redis can optionally be used to cache the etags and package data. Set
REDIS_URL to activate it.
Essentially the goal of the project is to quickly deliver current npm data even when npm is offline. In npm there are 2 main types of requests: package metadata and tarballs.
Package metadata mostly contains what versions of a package are available. These cannot be cached for very long since the package can be updated. By default, it is cached for 60 seconds. You can modify this with
CACHE_PACKAGE_TTL. Etags are also supported and cached to further speed up access.
The tarballs are the actual code and never change once they are uploaded (though they can be removed via unpublishing). These are downloaded one time from npmjs.org per package and version, stored locally or in S3 for future requests. These have a very long max-age header.
In the event npmjs.org is offline, npm-register will use the most recent package metadata that was requested from npmjs.org until it comes back online.
npm-register should support most npm commands. There are some exceptions, however:
If anything else doesn't work, please submit an issue so we can fix it, or at least note the missing functionality here.
npm-register uses an htpasswd file for authentication and stores tokens in S3. To set this up, first create an htpasswd file, then upload it to
/htpasswd in your S3 bucket or your local file system:
$ aws s3 cp s3://$AWS_S3_BUCKET/htpasswd ./htpasswd $ htpasswd -nB YOURUSERNAME >> ./htpasswd $ aws s3 cp ./htpasswd s3://$AWS_S3_BUCKET/htpasswd
Then you can login with npm. Note that the email is ignored by the server, but the CLI will force you to add one.
$ npm login --registry http://myregistry Username: jdxcode Password: Email: (this IS public) email@example.com $ npm whoami --registry http://myregistry jdxcode
This stores the credentials in
~/.npmrc. You can now use
npm publish to publish packages.
By default, all write endpoints (e.g. publish, unpublish) require authentication whereas read endpoints (e.g. install) don't. This default behaviour can be changed by using
NPM_REGISTER_AUTH_READ environment variables: use
true to enable authentication and
false to disable it.
Yarn doesn't follow HTTP redirects and so expects all URLs to be HTTPS by default. Pass
--always-https to ignore the protocol header and return all responses in a format Yarn
To run the tests:
Prerequisites for running the tests locally:
The s3 bucket needs read/write/delete access. Set the following env variables:
When running the test suite, you will need the following in
./tmp/htpasswd and in the root of your s3 bucket:
This sets up a test user with password 'test'.
When you have it running you should set the port number or url as the following env variable: