- 📚 Originally published on my Medium blog.
- 🔗 https://medium.com/p/c5ec915a430f
- 👏 Feel free to read it there, and clap/comment if you enjoyed it.
We are all apprentices in a craft where no one ever becomes a master.
I wanted to bring together some of the many patterns I’ve come to appreciate and actively implement in present and future projects.
If you have any comments or suggestions, please open an issue or a pull request with any changes that you would like to contribute! Thanks!
After initially publishing this article many years ago, many threads of discussion in channels such as Reddit have helped me adjust and tweak some of my explanations and stances on API design. I would like to thank all who have contributed to the discussion, and I hope this helps build this article into a more valuable resource for others!
If you’re going to develop an API for any client service, you’re going to want to prepare yourself for eventual change. This is best realised by providing a "version-namespace" for your RESTful API.
We do this with simply adding the version as a prefix to all URLs.
GET www.myservice.com/api/v1/posts
However, through studying other API implementations, I’ve grown to like a shorter URL style offered by accessing the API as part of a subdomain, and then dropping the /api
from the route; shorter and more concise is better.
GET api.myservice.com/v1/posts
It is important to consider that when placing your API into a different subdomain such as api.myservice.com
it will require implementing CORS for your backend if you plan to host your frontend site at www.myservice.com
and expect to use fetch requests without throwing No Access-Control-Allow-Origin header is present
errors.
Use HTTP methods such as:
GET
for fetching data.POST
for adding data.PUT
for updating data (as a whole object).PATCH
for updating data (with partial information for the object).DELETE
for deleting data.
I would like to add that I think PATCH
is great way to cut down the size of requests to change parts of bigger objects, but also that it fits well with commonly implemented auto-submit/auto-save fields.
A nice example of this is with Tumblr’s "Dashboard Settings" screen, where non-critical options about the user experience of the service can be edited and saved, per item, without the need of a final form submission button. It is simply a much more organic way to interact with the user’s preference data.
The "Saved" tag appears and then disappears shortly after modifying the option.
When building your routes you need to think of your endpoints as groups of resources from which you may read, add, edit and delete from and these actions are encapsulated as HTTP methods.
It makes semantic sense that you request many posts from /posts
.
And for goodness sake don’t consider mixing /post/all
with /post/:id
!
Although some resource names can't be pluralised, you should try to get as close to plural as you can.
// DO: plurals are consistent and make sense
GET /v1/posts/:id/attachments/:id/comments
// DON'T: is it just one comment? is it a form? etc.
GET /v1/post/:id/attachment/:id/comment
Query strings should be used for further filtering results beyond the initial grouping of a logical set offered by a relationship.
Aim to design endpoint paths that avoid unnecessary query string parameters as they are generally harder to read and to work with when compared to paths whose structure promotes an initial relationship-based filtering and grouping of such items the deeper it goes.
This
/posts/x/attachments
is better than/attachments?postId=x
.And this
/posts/x/attachments/y/comments
is so much better than/comments?postId=x&attachmentId=y
.
You should aim to keep your API as flat as possible and not crowd your resource paths. Allow yourself to provide root-level endpoints for all of your resources for when you eventually must update/delete them. For example, the case of a post having comments, allow GET /posts/:id/comments
to fetch the comments for a post based on that relationship, but also offer PATCH /comments/:id
to allow editing of comments without needing to track the id for that post for every single route.
- Longer paths for creating/fetching nested resources by relationship
- Shorter paths for updating/deleting resources based on their
id
.
When it comes to providing an endpoint to access to all of a user's own resources (e.g. all my own posts) you may end up with many ways to serve that information; it's up to you what best suits your application.
- Nest a
/posts
relationship under/me
withGET /me/posts
, or - Use the existing
/posts
endpoint but filter with query string,GET /posts?user=<id of self>
, or - Reuse
/posts
to show only your own posts, and expose public posts withGET /feed/posts
.
As hinted above, implement a GET /me
endpoint to deliver basic data about the user as distinguished through the Authorisation
header. This can include info about the user's permissions/scopes/groups/posts/sessions etc. that allow the client to show/hide elements and routes based on your permissions.
When it comes to providing endpoints for updating user preferences allow PATCH /me
to change those intrinsic values.
Pagination is necessary as your dataset grows and fetching potentially thousands of items from a collection is incredibly expensive for both the server and the client. One of the best ways to do handle pagination is by using query parameters such as from_<PROPERTY>
to indicate the beginning of your result set, with an optional limit
parameter that defaults to something like 20
with a maximum of 50
(depending on your use case). In your responses' meta
key you can provide information such as total
count, and a uniquely generated cursor
used for cursor-based pagination (used by most applications, particularly Google’s Places API and Twitter).
It is good to maintain a clean response structure for your APIs in the event that you need to add metadata (e.g. pagination information) to your responses. An easy way to do this is to envelope your application response data under a data
key, similar to GraphQL responses. From there you can freely add other keys, most likely an errors
key in which to return any errors that are not to be conflated with correct data
, and perhaps even a meta
key, in which you may include pagination information like total
and cursor
which, once again, are not to be conflated with data
items or indicative of any errors
.
If you choose to build your endpoints to allow for partial submission success (in which your transaction will still operate on and return correct data, but will also accumulate errors from incorrect data to respond afterwards) you will be able to return both successful response data with data
AND any problems with errors
. In the case of HTTP status codes, 207
seems the most appropriate.
Separating your responses' data types into data
, errors
, meta
, etc. helps remove validation headache when clients inspect your response, whereby you may easily check for the existence of an errors
key in the response rather than using keys like code
to differentiate an ITEM
from an ERROR
.
// DO: enveloped
{
data: ITEM or [ ITEM, ITEM, ... ],
meta: { total: 1000, cursor: "0794405882026854" },
errors: [
{ code: "MY_ERROR_CODE", extensions?: { ... } },
ERROR,
ERROR,
...
],
}
// DON'T non-enveloped
[
ITEM,
ITEM,
// ...
]
Most of the time you'll be communicating in JSON for both requests and responses.
If your API responses are in JSON, your server must include a Content-Type: application/json
header expressing the MIME-type in use.
For all other response formats such as CSVs/images/etc. you must define and return those respective MIME-types in your headers.
Most modern frameworks and tools for APIs and clients-alike will automatically handle these headers for you.
This header is set by the client making a request to the server and is used to inform the server as to what format it wants in return.
If you're only supporting JSON responses (instead of something like XML), then your API should return a 415
Unsupported Media Type HTTP status code for the resource that is trying to be accessed if the API is unable to respond in the format specified by the Accept
header.
I've seen some API's that use a /json
or /xml
suffix as part of a URL to access a resource, however this isn't exactly RESTful, and instead the Accept
header should be used.
When updating any resource through a PUT
or PATCH
it’s good practice to return the updated resource in response to a successful POST
, PUT
, or PATCH
request!
Support the 204 — No Content
response status code in cases where the request was successful but has no content to return. The envelope of the response, coupled with a 2XX
HTTP success code is enough to indicate a successful response without arbitrary "information".
DELETE /v1/posts/:id
// response - 204
{
"data": null
}
Because we are using HTTP methods, we should use HTTP status codes. Although a challenge here is to select a distinct slice of these codes, and then depend on response data to detail any response errors. Keeping a small set of codes helps you consume and handle errors consistently.
I like to use:
400
for when the requested information is incomplete or malformed.422
for when the requested information is okay, but invalid.404
for when everything is okay, but the resource doesn’t exist.409
for when a conflict of data exists, even with valid information.
401
for when an access token isn’t provided, or is invalid.403
for when an access token is valid, but requires more privileges.
200
for when everything is okay.204
for when everything is okay, but there’s no content to return.500
for when the server throws an error, completely unexpected.
Furthermore, returning responses after these errors is also very important. I want to consider not only the presentation of the status itself, but also a reason behind it.
In the case of trying to create a new account, imagine we provide an email
and password
. Of course we would like to have our client app prevent any requests with an invalid email, or password that is too short, but outsiders have as much access to the API as we do from our client app when it’s live.
- If the
email
field is missing, return a400
. - If the
password
field is too short, return a422
. - If the
email
field isn’t a valid email, return a422
. - If the
email
is already taken, return a409
.
"It’s much better to specify a more specific 4xx series code than just plain 400. I understand that you can put whatever you want in the response body to break down the error but codes are much easier to read at a glance." (source)
Now from these cases, two errors returned 422
s regardless of their reasons being different. This is why we need an error code, and maybe even an error description. It’s important to make a distinction between code and description as I intend to have code
as a machine consumable constant, and message
as a human consumable string that may change.
In the case of per-field errors, the presence of the field as a key in the error is enough of a "code" to indicate that it is a target of a validation error.
For returning those per field errors, it may be returned as:
POST /v1/register
// request
{
"email": "end@@user.comx"
"password": "abc"
}
// response - 422
{
"errors": [{
"code": "FIELDS_VALIDATION_ERROR",
"extensions": {
"email": "Invalid email address.",
"password": "Password too short."
}
}]
}
And for returning operational validation errors:
POST /v1/register
// request
{
"email": "end@user.com",
"password": "password"
}
// response - 409
{
"error": {
"code": "EMAIL_ALREADY_EXISTS",
"message": "An account already exists with this email."
}
}
This way, your fetch logic watches out for non-200 errors, and can then straight-up check the error
key from the response and then compare it to any further logic in the client app.
The message
can act as a fallback human-readable error message to help understand the request when developing, and also in the case an appropriate localisation string implementation cannot be used.
Modern stateless, RESTful APIs implement authentication with tokens most commonly provided through the Authorization
header (or even an access_token
query param).
Originally I thought that issuing JWTs for regular API requests was a great way to handle authentication—until I wanted to invalidate those tokens.
In my last revision of this post (and detailed in a separate post) I offered a way for JWTs to be reissued through an additionally stored client secret "Refresh Token" (RT) which was to be exchanged for new JWTs. However in order to expire these JWTs they each contained a reference to the issuing RT so if the RT was invalided/deleted so would the JWT. However this mechanism defeats the statelessness of the JWT itself...
My solution now is to simply use a /sessions
resource endpoint to exchange login credentials for a single unique session token (using uuid4
) which is hashed and stored as a database row. Just like many moderns apps, the token doesn't need to be reissued unless there is a long period of inactivity (similar to session timeout, but to the scale of weeks). After initial authentication, every future request bumps the life of the token in a self-extending manner as long as it hasn't expired.
A normal login process would look like:
- Receive email/password combination with
POST /sessions
, treatingsessions
as just another resource. - Check the email/password-hash against the database.
- Create a new
session
database row that contains a hasheduuid4()
as atoken
. - Return the non-hashed
token
string to the client.
In this flow tokens don't need to be explicitly renewed or reissued. That's because the API extends the life of the token if its still valid every request, saving regular users from ever having a session expire for them.
Whenever a token is received by the API i.e. through an Authorization
header:
- Receive the
token
i.e. from theAuthorization
header. - Compare against the
token
's hash, if there is no matchingsession
row, raise an authentication error. - Check the
updated_at
property of thesession
, if now if greater thanupdated_at + session_life
the session is considered expired, delete thesession
row, raise an authentication error. - If it exists and is still valid from
updated_at
time, set theupdated_at
time tonow()
to renew the token.
Because all sessions are tracked as database rows mapped to a user, a user can see all their active sessions similar to Facebook's account security sessions view. You can also chose to include any associated metadata you have chosen to collected when initially creating a session such as the browser's User Agent, IP address etc. And
Fetching all your sessions is as simple as:
GET /sessions
to return all sessions associated with your user via theAuthorization
header.
And because you have handles to your sessions you can terminate them to invalidate unauthorised or unwanted access to your account. And logging out would simply be terminating the client's session and purging the session from the client.
- Receive the
token
as part of aDELETE /sessions/id
request. - Compare against the
token
's hash, delete the matching session row.
After doing a lot of research into password rules, I’ve come to agree that password rules are bullshit and are part of NIST’s "don’ts", especially considering that password composition rules help narrow down valid passwords based on their validity rules.
I’ve collated some of the best points (from the above links) for password handling:
- Only enforce a minimum unicode password length (min 8-10).
- Check against common passwords ("password12345")
- Check for basic entropy (don’t allow "aaaaaaaaaaaaa").
- Don’t use password composing rules (at least one "!@#$%&").
- Don’t use password hints (rhymes with "assword").
- Don’t use knowledge-based authentication ("security" questions).
- Don’t expire passwords without reason.
- Don’t use SMS for two-factor authentication.
- Use a password salt of 32-bits or more.
These "don’ts" should make password validation much easier!
Through developing with AWS, it been necessary to provide a way to output a simple response that can indicate that the API instance is alive and does not need to be restarted. It’s also useful for easily checking what version of the API is on any machine at any time, without authentication.
GET /v1
// response - 200
{
"status": "running",
"version": "fdb1d5e"
}
I provide status
and version
(which refers to the git commit ref of the API at the time it was built). It’s also worth mentioning that this value is not derived from an active .git
repo being bundled with the APIs container for EC2. Instead, it is read (and stored in memory) on initialisation from a version.txt
file (which is generated from the build process), and defaults to __UNKNOWN__
in case of a read error, or the file does not exist.