From Newsgroup: rec.sport.rowing
The Drupal Commerce Log module provides activity logs for Commerce entities. Administrative users can view these logs through the UI to help them manage these entities. Currently, the module only includes built-in support for logging activity related to Order entities. However, developers will find that the module can be easily extended to log activity for other entities.
drupal download logs
Download
https://t.co/uwbW0G3Ngo
The next two logs were created when the order was placed. All of the default Order state transition events are logged: Placed, Validated, Fulfilled, and Canceled. For more information on state transition events, see the State machine documentation.
Log files track and record your site's activity to help you find, debug, and isolate current or potential problems on your site. Each environment (Multidev, Dev, Test, and Live) has individual log files, which can be obtained via SFTP. Application-level logs can be accessed directly through Drupal. You can use New Relic Performance Monitoring in conjunction with logs to help diagnose and fix errors and performance bottlenecks.
Log files are destroyed during appserver migration (a standard part of regular platform maintenance). Log deletion occurs because the log files are appserver-specific. Consider automating the collection of logs regularly to maintain historical log data.
Rotated log files are archived within the /logs directory on application containers and database servers. The directory might contain sub-directories for services like Nginx and PHP, or it may log files directly in /logs For example:
Of course, if you actually want to use the Statistics module, then make sure that Discard access logs older than is set to something other than 'never', so that the accesslog table doesn't grow indefinitely.
Drupal's default installation profile come with (Database Logging)dblog enabled, which facilitates you to see the logs on the websiteitself. However for (high traffic) live sites it is not a good idea tokeep it enabled.Luckily there are alternatives:
The focus of this presentation will be on how easy it is to start. We will cover how to run the ELK stack on your development machine with just a few commands - using the power of docker - and analyse your apache, Drupal and php logs.
OK, do you see it? The watchdog() function doesn't actually log anything at all. It may seem bonkers, but it's actually very well designed. The whole point of the watchdog() function is to provide a single function that every Drupal module can call as needed, regardless of where logs are being stored, how logs are being processed, or if logging is actually happening at all.
The second parameter of variable_get() is the default value to return if the variable doesn't exist. The best part about this code is you don't need to define this variable when you push your module. When you encounter a production issue and you need the logs simply use the magic of Drush aliases to set this variable to enable logging. No code push needed! When it's time to turn off the logging simple use Drush to disable the variable.
Events that occur within an app container are logged within that container.The logs can be written to, but you should do so only with standard logging mechanisms.If your app has its own logging mechanism, use it to write to a dedicated logs mount.
This code will request the results of www.google.com, and place them in the $result variable. In the case that the request failed for some reason, the system logs the result of $error->getMessage() to the Drupal log.
I would recommend looking through some system logs to ensure that there was no access to the server, and ensure that not only the CMS (Drupal) is up to date, but all of the plugins, and themes as well. If you do some searching around, you may find that even an up to date plugin may have vulnerabilities.
If Apache is unable to open the configuration or the log file, check that the owner of those files is the same user account that installed Apache and that it has write permissions on logs and read permissions on the configuration file. If this is not the case, you will see these errors:
The Amazon EC2 instances in your Elastic Beanstalk environment generate logs that you can view to troubleshoot issues with your application or configuration files. Logs created by the web server, application server, Elastic Beanstalk platform scripts, and AWS CloudFormation are stored locally on individual instances. You can easily retrieve them by using the environment management console or the EB CLI. You can also configure your environment to stream logs to Amazon CloudWatch Logs in real time.
Bundle logs are full logs for a wider range of log files, including logs from yum and cron and several logs from AWS CloudFormation. When you request bundle logs, an instance in your environment packages the full log files into a ZIP archive and uploads it to Amazon S3.
To upload rotated logs to Amazon S3, the instances in your environment must have an instance profile with permission to write to your Elastic Beanstalk Amazon S3 bucket. These permissions are included in the default instance profile that Elastic Beanstalk prompts you to create when you launch an environment in the Elastic Beanstalk console for the first time.
Elastic Beanstalk stores tail and bundle logs in an Amazon S3 bucket, and generates a presigned Amazon S3 URL that you can use to access your logs. Elastic Beanstalk deletes the files from Amazon S3 after a duration of 15 minutes.
To persist logs, you can configure your environment to publish logs to Amazon S3 automatically after they are rotated. To enable log rotation to Amazon S3, follow the procedure in Configuring instance log viewing. Instances in your environment will attempt to upload logs that have been rotated once per hour.
If your application generates logs in a location that isn't part of the default configuration for your environment's platform, you can extend the default configuration by using configuration files (.ebextensions). You can add your application's log files to tail logs, bundle logs, or log rotation.
When you request tail or bundle logs from your environment, or when instances upload rotated logs, they're stored in your Elastic Beanstalk bucket in Amazon S3. Elastic Beanstalk creates a bucket named elasticbeanstalk-region-account-id for each AWS Region in which you create environments. Within this bucket, logs are stored under the path resources/environments/logs/logtype/environment-id/instance-id.
On Linux platforms, Elastic Beanstalk uses logrotate to rotate logs periodically. If configured, after a log is rotated locally, the log rotation task picks it up and uploads it to Amazon S3. Logs that are rotated locally don't appear in tail or bundle logs by default.
Elastic Beanstalk uses files in subfolders of /opt/elasticbeanstalk/tasks (Linux) or C:\Program Files\Amazon\ElasticBeanstalk\config (Windows Server) on the Amazon EC2 instance to configure tasks for tail logs, bundle logs, and log rotation.
Use the files section to add configuration files to the tasks that you want to modify. For example, the following configuration text adds a log configuration file to each instance in your environment. This log configuration file, cloud-init.conf, adds /var/log/cloud-init.log to tail logs.
On Linux platforms, you can also use wildcard characters in log task configurations. This configuration file adds all files with the .log file name extension from the log folder in the application root to bundle logs.
Much like extending tail logs and bundle logs, you can extend log rotation using a configuration file. Whenever Elastic Beanstalk rotates its own logs and uploads them to Amazon S3, it also rotates and uploads your additional logs. Log rotation extension behaves differently depending on the platform's operating system. The following sections describe the two cases.
As explained in Log rotation settings on Linux, Elastic Beanstalk uses logrotate to rotate logs on Linux platforms. When you configure your application's log files for log rotation, the application doesn't need to create copies of log files. Elastic Beanstalk configures logrotate to create a copy of your application's log files for each rotation. Therefore, the application must keep log files unlocked when it isn't actively writing to them.
You can configure your environment to stream logs to Amazon CloudWatch Logs in the Elastic Beanstalk console or by using configuration options. With CloudWatch Logs, each instance in your environment streams logs to log groups that you can configure to be retained for weeks or years, even after your environment is terminated.
You can configure log streaming in the Elastic Beanstalk console either during environment creation or for an existing environment. In the following example, logs are saved for up to seven days, even when the environment is terminated.
If logs are missing, it could be due to incorrect permissions or configuration issues. Ensure that your logging configuration in settings.php is correct and that Drupal has the necessary permissions to write to log files or the database.
Similar to drupal 7, Drupal 8 also provide the logging functionality but little different to use it because the drupal 8 is symphony base architecture, so can use the logger service of drupal 8 to reload your logs on dblog
f5d0e4f075
--- Synchronet 3.21a-Linux NewsLink 1.2