Business Security Test March-April 2020 – Factsheet
|Test Period||March - April 2020|
|Number of Testcases||403 Real-World
1,192 Malware Protection
|Online with cloud connectivity|
|False Alarm Test included|
This is a short fact sheet for our Business Main-Test Series, containing the results of the Business Malware Protection Test (March) and Business Real-World Protection Test (March–April). The full report, including the Performance Test and product reviews, will be released in July.
To be certified in December as an “Approved Business Product” by AV-Comparatives, the tested products must score at least 90% in the Malware Protection Test, and at least 90% in the overall Real-World Protection Test (i.e. over the course of 4 months), with zero false alarms on common business software. Tested products must also avoid major performance issues and have fixed all reported bugs in order to gain certification.
Please note that the results of the Business Main-Test Series cannot be compared with the results of the Consumer Main-Test Series, as the tests are done at different times, with different sets, different settings, etc.
The following products were tested under Windows 10 1909 64-bit and are included in this factsheet:
Information about additional third-party engines/signatures used by some of the products: Acronis, Cisco, Cybereason, FireEye, G DATA and VIPRE use the Bitdefender engine (in addition to their own protection features). VMware uses the Avira engine (in addition to their own protection features). G DATA’s OutbreakShield is based on Cyren.
In business environments, and with business products in general, it is usual for products to be configured by the system administrator, in accordance with vendor’s guidelines, and so we invited all vendors to configure their respective products.
Only a few vendors provide their products with optimal default settings which are ready to use, and did therefore not change any settings. Cloud and PUA detection have been activated in all products. We currently do not include any PUA in our malware tests.
Please keep in mind that the results reached in the Enterprise Main-Test Series were only achieved by applying the respective product configurations described here. Any setting listed here as enabled might be disabled in your environment, and vice versa. This influences the protection rates, false alarm rates and system impact. The applied settings are used across all our Enterprise Tests over the year. That is to say, we do not allow a vendor to change settings depending on the test. Otherwise, vendors could e.g. configure their respective products for maximum protection in the protection tests (which would reduce performance and increase false alarms), and maximum speed in the performance tests (thus reducing protection and false alarms). Please not that some enterprise products have all their protection features disabled by default, so the admin has to configure the product to get any protection.
Below we have listed relevant deviations from default settings (i.e. setting changes applied by the vendors):
Bitdefender: “Sandbox Analyzer” and “Scan SSL” enabled; “HyperDetect”, “Device Sensor” and “EDR Sensor” disabled.
Cisco: everything enabled and set to Block.
CrowdStrike: everything enabled and set to maximum, i.e. “Extra Aggressive”. “Unknown Detection-Related Executables” and “Unknown Executables” disabled.
Cybereason: “Anti-Malware” enabled; “Signatures mode” set to “Disinfect”; “Behavioral document protection” enabled; “Artificial intelligence” set to “Aggressive”; “Exploit protection”, “PowerShell and .NET”, “Anti-Ransomware” and “App Control” enabled and set to “Prevent”; all “Collection features” enabled; “Scan archives on access” enabled.
Elastic: “Malware” and “Process Injection” protections enabled; “Blacklist”, “Credential Access”, “Exploit” and “Ransomware” protections, as well as all “Adversary Behaviors” disabled.
ESET: All “Real-Time & Machine Learning Protection” settings set to “Aggressive”.
FireEye: “Real-Time Indicator Detection” disabled, “Exploit Guard” and “Malware Protection” enabled.
Fortinet: All “AntiVirus Protection” settings enabled and set to “Block”. Additionally, “Anti-Exploit”, “Cloud Based Malware Detection”, “Advanced Heuristic”, “FortiGuard Analytics”, FortiSandbox’s “Sandbox Detection”, “Web Filter”, “Application Firewall”, “Detect and Block Exploits & Botnets” and “FortiEDR” were all enabled; “Exclude Files from Trusted Sources” for “Sandbox Detection” enabled.
G DATA: “Exploit Protection”, “Anti-Ransomware” and “BankGuard” enabled; “BEAST Behavior Monitoring” set to “Pause Program and Quarantine”.
Kaspersky: “Adaptive Anomaly Control” disabled.
Microsoft: Cloud protection level set to “High”, Cloud-delivered protection set to “Advanced”. Google Chrome extension “Windows Defender Browser Protection” installed and enabled.
Sophos: All options in “Active Adversary Mitigations” enabled. “Web Control” and “Protect against data loss” disabled.
SparkCognition: all “Policy Settings” and all “Attack Vectors” settings enabled and set to “Aggressive”.
VMware: policy set to “Advanced”.
Acronis, Avast, K7, Panda, VIPRE: default settings.
Real-World Protection Test (March-April)
This fact sheet gives a brief overview of the results of the Business Real-World Protection Test run in March and April 2020. The overall business product reports (each covering four months) will be released in July and December. For more information about this Real-World Protection Test, please read the details available at https://www.av-comparatives.org. The results are based on a test set consisting of 403 test cases (such as malicious URLs), tested from the beginning of March till the end of April.
|Blocked||User dependent||Compromised||PROTECTION RATE
[Blocked % + (User dependent %)/2]*
|Cisco, G Data||400||–||3||99.3%||1|
* User-dependent cases are given half credit. For example, if a program blocks 80% by itself, and another 20% of cases are user-dependent, we give half credit for the 20%, i.e. 10%, so it gets 90% altogether.
Malware Protection Test (March)
The Malware Protection Test assesses a security program’s ability to protect a system against infection by malicious files before, during or after execution. The methodology used for each product tested is as follows. Prior to execution, all the test samples are subjected to on-access scans (if this feature is available) by the security program (e.g. while copying the files over the network). Any samples that have not been detected by the on-access scanner are then executed on the test system, with Internet/cloud access available, to allow e.g. behavioral detection features to come into play. If a product does not prevent or reverse all the changes made by a particular malware sample within a given time period, that test case is considered to be a miss. For this test, 1,192 recent malware samples were used.
False positive (false alarm) test with common business software
A false alarm test done with common business software was also performed. All tested products had zero false alarms on common business software.
The following chart shows the results of the Business Malware Protection Test:
|Malware Protection Rate||False Alarms on common business software|
|Cisco, K7, Microsoft, VMware
|Bitdefender, ESET, G Data, Panda||99.9%||0|
* A SparkCognition product issue was uncovered during the Malware Protection Test which led to some missed detections. The bug has now been fixed.
In order to better evaluate the products’ detection accuracy and file detection capabilities (ability to distinguish good files from malicious files), we also performed a false alarm test on non-business software and uncommon files. This is provided mainly just as additional information, especially for organisations which often use uncommon non-business software or their own self-developed software. The results do not affect the overall test score or the Approved Business Product award. The false alarms found were promptly fixed by the respective vendors.
|FP rate on non-business software|
|Acronis, Avast, Bitdefender, Cisco, ESET, Fortinet, G Data, Kaspersky, Sophos||Very low|
|Cybereason, FireEye, SparkCognition, Microsoft||Low|
|Elastic, Vipre, VMware||Medium|
Copyright and Disclaimer
This publication is Copyright © 2020 by AV-Comparatives ®. Any use of the results, etc. in whole or in part, is ONLY permitted after the explicit written agreement of the management board of AV-Comparatives prior to any publication. AV-Comparatives and its testers cannot be held liable for any damage or loss, which might occur as result of, or in connection with, the use of the information provided in this paper. We take every possible care to ensure the correctness of the basic data, but a liability for the correctness of the test results cannot be taken by any representative of AV-Comparatives. We do not give any guarantee of the correctness, completeness, or suitability for a specific purpose of any of the information/content provided at any given time. No one else involved in creating, producing or delivering test results shall be liable for any indirect, special or consequential damage, or loss of profits, arising out of, or related to, the use or inability to use, the services provided by the website, test documents or any related data.
For more information about AV-Comparatives and the testing methodologies, please visit our website.