One of the biggest challenges when developing applications for the web is to understand how they need to be fine-tuned when releasing them into a production environment. This is no exception for Java Enterprise applications deployed on a Payara Server installation.
Running a Payara Server setup is simple: download the current distribution suited for your needs (full, web); head to the /bin folder and start the default domain (domain1)! However, keep in mind that this default domain is tailored for development purposes (a trait inherited from GlassFish Server Open Source). When developing a web application, it’s better to quickly code features, deploy them quickly, test it, un-deploy (or redeploy) it, and continue with the next set of features until a stable state is reached.
(last updated 06/04/2021)
Nowadays, the concept of microservices is more than a simple novelty. With the advent of DevOps and the boom of container technologies and deployment automation tools, microservices are changing the way developers structure their applications. With this article, our intention is to illustrate that microservices are a valid option for Java/Jakarta EE developers and how Payara Micro is a robust platform to reach that goal.
Starting this year, customers that come on board with our support services also have access to commercial OpenJDK support included, thanks to the partnership between Payara Services and Azul Systems' Enterprise Support! If you are interested in Payara Enterprise or our Migration & Project Support but are hesitating and have some doubts about what value this service brings to your organization and environment, this article may help dispel them and give you much needed decision-making clarity.
As part of release 220.127.116.11, new notifier integrations were developed for Payara Server for the New Relic and DataDog application performance monitoring (APM) services. Both services allow the gathering of JVM statistics, HTTP metrics and support the use of notification for critical events in the server lifecycle management. In this era of cloud services, performance monitoring is an integral part of the IT infrastructure for any organization, which is the reason integration with these services has been brought to Payara Server. This article will show how to correctly set up these notifiers to that purpose.
A crucial component in many web applications is the use of a database. The chances of using JDBC directly or indirectly to access a relational database (through JPA or other ORM frameworks) on these applications are quite high. A common problem when dealing with relational databases is dealing with SQL queries or statements that take too much time to resolve, thus causing your application to be considered slow in producing results to the user. It's usually better to detect these issues preemptively, before sub-par performance damages your application, and Payara Server has you covered!
One of the core steps in every continuous integration process is running integration tests for your application. Unlike vanilla unit tests, integration tests allow you to assess the state of your applications or systems by testing all of its components together (modules, databases, messaging, etc.) and verifying that they work correctly as a whole unit. Needless to say, integration tests are more complex that simple unit tests, have a larger footprint, take more time and are usually saved to test full releases or major changes to implementations.
As previously reported on this blog, the Request Tracing Service was improved drastically in release 18.104.22.168 and implemented the configuration of a historic trace record storing for increased productivity purposes. In addition to these changes, we also made the configuration on the Request Tracing Service in Payara Micro for the same release. These changes to Payara Micro make it simpler to configure the Request Tracing Service when starting a new instance!
Following up from the first part of the Security Auditing article, where we covered the audit logging, in this part we will focus on creating a custom audit module.