Mongodb Check if Database Exists Node Js
In this article, we'll explore the Mongoose library for MongoDB. Mongoose is a Object Data Modeling (ODM) library for MongoDB distributed as an npm package. We'll compare and contrast Mongoose to using the native MongoDB Node.js driver together with MongoDB Schema Validation.
We'll see how the MongoDB Schema Validation helps us enforce a database schema, while still allowing for great flexibility when needed. Finally, we'll see if the additional features that Mongoose provides are worth the overhead of introducing a third-party library into our applications.
#What is Mongoose?
Mongoose is a Node.js-based Object Data Modeling (ODM) library for MongoDB. It is akin to an Object Relational Mapper (ORM) such as SQLAlchemy for traditional SQL databases. The problem that Mongoose aims to solve is allowing developers to enforce a specific schema at the application layer. In addition to enforcing a schema, Mongoose also offers a variety of hooks, model validation, and other features aimed at making it easier to work with MongoDB.
#What is MongoDB Schema Validation?
MongoDB Schema Validation makes it possible to easily enforce a schema against your MongoDB database, while maintaining a high degree of flexibility, giving you the best of both worlds. In the past, the only way to enforce a schema against a MongoDB collection was to do it at the application level using an ODM like Mongoose, but that posed significant challenges for developers.
#Getting Started
If you want to follow along with this tutorial and play around with schema validations but don't have a MongoDB instance set up, you can set up a free MongoDB Atlas cluster here.
#Object Data Modeling in MongoDB
A huge benefit of using a NoSQL database like MongoDB is that you are not constrained to a rigid data model. You can add or remove fields, nest data multiple layers deep, and have a truly flexible data model that meets your needs today and can adapt to your ever-changing needs tomorrow. But being too flexible can also be a challenge. If there is no consensus on what the data model should look like, and every document in a collection contains vastly different fields, you're going to have a bad time.
#Mongoose Schema and Model
On one end of the spectrum, we have ODM's like Mongoose, which from the get-go force us into a semi-rigid schema. With Mongoose, you would define a Schema
object in your application code that maps to a collection in your MongoDB database. The Schema
object defines the structure of the documents in your collection. Then, you need to create a Model
object out of the schema. The model is used to interact with the collection.
For example, let's say we're building a blog and want to represent a blog post. We would first define a schema and then create an accompanying Mongoose model:
1 const blog = new Schema ({ 2 title: String, 3 slug: String, 4 published: Boolean, 5 content: String, 6 tags: [String], 7 comments: [{ 8 user: String, 9 content: String, 10 votes: Number 11 }] 12 }); 13 14 const Blog = mongoose .model ('Blog', blog) ;
#Executing Operations on MongoDB with Mongoose
Once we have a Mongoose model defined, we could run queries for fetching, updating, and deleting data against a MongoDB collection that aligns with the Mongoose model. With the above model, we could do things like:
1 // Create a new blog post 2 const article = new Blog ({ 3 title: 'Awesome Post!', 4 slug: 'awesome-post', 5 published: true, 6 content: 'This is the best post ever', 7 tags: ['featured', 'announcement'], 8 }); 9 10 // Insert the article in our MongoDB database 11 article .save ( ); 12 13 // Find a single blog post 14 Blog .findOne ( { } , (err, post) => { 15 console.log ( post); 16 }) ;
#Mongoose vs MongoDB Node.js Driver: A Comparison
The benefit of using Mongoose is that we have a schema to work against in our application code and an explicit relationship between our MongoDB documents and the Mongoose models within our application. The downside is that we can only create blog posts and they have to follow the above defined schema. If we change our Mongoose schema, we are changing the relationship completely, and if you're going through rapid development, this can greatly slow you down.
The other downside is that this relationship between the schema and model only exists within the confines of our Node.js application. Our MongoDB database is not aware of the relationship, it just inserts or retrieves data it is asked for without any sort of validation. In the event that we used a different programming language to interact with our database, all the constraints and models we defined in Mongoose would be worthless.
On the other hand, if we decided to use just the MongoDB Node.js driver, we could run queries against any collection in our database, or create new ones on the fly. The MongoDB Node.js driver does not have concepts of object data modeling or mapping.
We simply write queries against the database and collection we wish to work with to accomplish the business goals. If we wanted to insert a new blog post in our collection, we could simply execute a command like so:
1 db .collection ('posts') .insertOne ({ 2 title: 'Better Post!', 3 slug: 'a-better-post', 4 published: true, 5 author: 'Ado Kukic', 6 content: 'This is an even better post', 7 tags: ['featured'], 8 }) ;
This insertOne()
operation would run just fine using the Node.js Driver. If we tried to save this data using our Mongoose Blog
model, it would fail, because we don't have an author
property defined in our Blog Mongoose model.
Just because the Node.js driver doesn't have the concept of a model, does not mean we couldn't create models to represent our MongoDB data at the application level. We could just as easily create a generic model or use a library such as objectmodel. We could create a Blog
model like so:
1 function Blog(post) { 2 this.title = post .title; 3 this.slug = post .slug; 4 ... 5 }
We could then use this model in conjunction with our MongoDB Node.js driver, giving us both the flexibility of using the model, but not being constrained by it.
1 db .collection ('posts') .findOne ( { }) .then ((err, post) => { 2 let article = new Blog ( post); 3 }) ;
In this scenario, our MongoDB database is still blissfully unaware of our Blog model at the application level, but our developers can work with it, add specific methods and helpers to the model, and would know that this model is only meant to be used within the confines of our Node.js application. Next, let's explore schema validation.
#Adding Schema Validation
We can choose between two different ways of adding schema validation to our MongoDB collections. The first is to use application-level validators, which are defined in the Mongoose schemas. The second is to use MongoDB schema validation, which is defined in the MongoDB collection itself. The huge difference is that native MongoDB schema validation is applied at the database level. Let's see why that matters by exploring both methods.
#Schema Validation with Mongoose
When it comes to schema validation, Mongoose enforces it at the application layer as we've seen in the previous section. It does this in two ways.
First, by defining our model, we are explicitly telling our Node.js application what fields and data types we'll allow to be inserted into a specific collection. For example, our Mongoose Blog schema defines a title
property of type String
. If we were to try and insert a blog post with a title
property that was an array, it would fail. Anything outside of the defined fields, will also not be inserted in the database.
Second, we further validate that the data in the defined fields matches our defined set of criteria. For example, we can expand on our Blog model by adding specific validators such as requiring certain fields, ensuring a minimum or maximum length for a specific field, or coming up with our custom logic even. Let's see how this looks with Mongoose. In our code we would simply expand on the property and add our validators:
1 const blog = new Schema ({ 2 title: { 3 type: String, 4 required: true, 5 }, 6 slug: { 7 type: String, 8 required: true, 9 }, 10 published: Boolean, 11 content: { 12 type: String, 13 required: true, 14 minlength: 250 15 }, 16 ... 17 }); 18 19 const Blog = mongoose .model ('Blog', blog) ;
Mongoose takes care of model definition and schema validation in one fell swoop. The downside though is still the same. These rules only apply at the application layer and MongoDB itself is none the wiser.
The MongoDB Node.js driver itself does not have mechanisms for inserting or managing validations, and it shouldn't. We can define schema validation rules for our MongoDB database using the MongoDB Shell or Compass.
We can create a schema validation when creating our collection or after the fact on an existing collection. Since we've been working with this blog idea as our example, we'll add our schema validations to it. I will use Compass and MongoDB Atlas. For a great resource on how to programmatically add schema validations, check out this series.
If you want to follow along with this tutorial and play around with schema validations but don't have a MongoDB instance set up, you can set up a free MongoDB Atlas cluster here.
Create a collection called posts
and let's insert our two documents that we've been working with. The documents are:
1 [ {"title":"Better Post!","slug":"a-better-post","published":true,"author":"Ado Kukic","content":"This is an even better post","tags": ["featured"]} , {"_id": {"$oid":"5e714da7f3a665d9804e6506"} ,"title":"Awesome Post","slug":"awesome-post","published":true,"content":"This is an awesome post","tags": ["featured","announcement"]}]
Now, within the Compass UI, I will navigate to the Validation tab. As expected, there are currently no validation rules in place, meaning our database will accept any document as long as it is valid BSON. Hit the Add a Rule button and you'll see a user interface for creating your own validation rules.
By default, there are no rules, so any document will be marked as passing. Let's add a rule to require the author
property. It will look like this:
1 { 2 $jsonSchema: { 3 bsonType: "object", 4 required: [ "author" ] 5 } 6 }
Now we'll see that our initial post, that does not have an author
field has failed validation, while the post that does have the author
field is good to go.
We can go further and add validations to individual fields as well. Say for SEO purposes we wanted all the titles of the blog posts to be a minimum of 20 characters and have a maximum length of 80 characters. We can represent that like this:
1 { 2 $jsonSchema: { 3 bsonType: "object", 4 required: [ "tags" ], 5 properties: { 6 title: { 7 type: "string", 8 minLength: 20, 9 maxLength: 80 10 } 11 } 12 } 13 }
Now if we try to insert a document into our posts
collection either via the Node.js Driver or via Compass, we will get an error.
There are many more rules and validations you can add. Check out the full list here. For a more advanced guided approach, check out the articles on schema validation with arrays and dependencies.
#Expanding on Schema Validation
With Mongoose, our data model and schema are the basis for our interactions with MongoDB. MongoDB itself is not aware of any of these constraints, Mongoose takes the role of judge, jury, and executioner on what queries can be executed and what happens with them.
But with MongoDB native schema validation, we have additional flexibility. When we implement a schema, validation on existing documents does not happen automatically. Validation is only done on updates and inserts. If we wanted to leave existing documents alone though, we could change the validationLevel
to only validate new documents inserted in the database.
Additionally, with schema validations done at the MongoDB database level, we can choose to still insert documents that fail validation. The validationAction
option allows us to determine what happens if a query fails validation. By default, it is set to error
, but we can change it to warn
if we want the insert to still occur. Now instead of an insert or update erroring out, it would simply warn the user that the operation failed validation.
And finally, if we needed to, we can bypass document validation altogether by passing the bypassDocumentValidation
option with our query. To show you how this works, let's say we wanted to insert just a title
in our posts
collection and we didn't want any other data. If we tried to just do this...
1 db .collection ('posts') .insertOne ( { title: 'Awesome' }) ;
... we would get an error saying that document validation failed. But if we wanted to skip document validation for this insert, we would simply do this:
1 db .collection ('posts') .insertOne( 2 { title: 'Awesome' }, 3 { bypassDocumentValidation: true } 4 ) ;
This would not be possible with Mongoose. MongoDB schema validation is more in line with the entire philosophy of MongoDB where the focus is on a flexible design schema that is quickly and easily adaptable to your use cases.
#Populate and Lookup
The final area where I would like to compare Mongoose and the Node.js MongoDB driver is its support for pseudo-joins. Both Mongoose and the native Node.js driver support the ability to combine documents from multiple collections in the same database, similar to a join in traditional relational databases.
The Mongoose approach is called Populate. It allows developers to create data models that can reference each other and then, with a simple API, request data from multiple collections. For our example, let's expand on the blog post and add a new collection for users.
1 const user = new Schema ({ 2 name: String, 3 email: String 4 }); 5 6 const blog = new Schema ({ 7 title: String, 8 slug: String, 9 published: Boolean, 10 content: String, 11 tags: [String], 12 comments: [{ 13 user: { Schema .Types .ObjectId , ref: 'User' }, 14 content: String, 15 votes: Number 16 }] 17 }); 18 19 const User = mongoose .model ('User', user); 20 const Blog = mongoose .model ('Blog', blog) ;
What we did above was we created a new model and schema to represent users leaving comments on blog posts. When a user leaves a comment, instead of storing information on them, we would just store that user's _id
. So, an update operation to add a new comment to our post may look something like this:
1 Blog .updateOne ({ 2 comments: [ { user: "12345", content: "Great Post!!!" }] 3 }) ;
This is assuming that we have a user in our User
collection with the _id
of 12345
. Now, if we wanted to populate our user
property when we do a query—and instead of just returning the _id
return the entire document—we could do:
1 Blog. 2 findOne ( {}). 3 populate ('comments.user'). 4 exec ( function (err, post) { 5 console.log ( post .comments [0] .user .name ) // Name of user for 1st comment 6 }) ;
Populate coupled with Mongoose data modeling can be very powerful, especially if you're coming from a relational database background. The drawback though is the amount of magic going on under the hood to make this happen. Mongoose would make two separate queries to accomplish this task and if you're joining multiple collections, operations can quickly slow down.
The other issue is that the populate concept only exists at the application layer. So while this does work, relying on it for your database management can come back to bite you in the future.
MongoDB as of version 3.2 introduced a new operation called $lookup
that allows to developers to essentially do a left outer join on collections within a single MongoDB database. If we wanted to populate the user information using the Node.js driver, we could create an aggregation pipeline to do it. Our starting point using the $lookup
operator could look like this:
1 db .collection ('posts') .aggregate ([ 2 { 3 '$lookup': { 4 'from': 'users', 5 'localField': 'comments.user', 6 'foreignField': '_id', 7 'as': 'users' 8 } 9 } , {} 10 ] , (err, post) => { 11 console.log ( post .users ) ; //This would contain an array of users 12 }) ;
We could further create an additional step in our aggregation pipeline to replace the user information in the comments
field with the users data, but that's a bit out of the scope of this article. If you wish to learn more about how aggregation pipelines work with MongoDB, check out the aggregation docs.
#Final Thoughts: Do I Really Need Mongoose?
Both Mongoose and the MongoDB Node.js driver support similar functionality. While Mongoose does make MongoDB development familiar to someone who may be completely new, it does perform a lot of magic under the hood that could have unintended consequences in the future.
I personally believe that you don't need an ODM to be successful with MongoDB. I am also not a huge fan of ORMs in the relational database world. While they make initial dive into a technology feel familiar, they abstract away a lot of the power of a database.
Developers have a lot of choices to make when it comes to building applications. In this article, we looked at the differences between using an ODM versus the native driver and showed that the difference between the two is not that big. Using an ODM like Mongoose can make development feel familiar but forces you into a rigid design, which is an anti-pattern when considering building with MongoDB.
The MongoDB Node.js driver works natively with your MongoDB database to give you the best and most flexible development experience. It allows the database to do what it's best at while allowing your application to focus on what it's best at, and that's probably not managing data models.
- MongoDB
- Atlas
- JavaScript
- ...
Mongodb Check if Database Exists Node Js
Source: https://www.mongodb.com/developer/article/mongoose-versus-nodejs-driver/