Peek Inside the Closed World: Evaluating Autoencoder-Based Detection of DDoS to Cloud. Guo, H., Fan, X., Cao, A., Outhred, G., & Heidemann, J. Technical Report arXiv:1912.05590v2 [cs.NI], arXiv, December, 2019. Paper abstract bibtex Machine-learning-based anomaly detection (ML-based AD) has been successful at detecting DDoS events in the lab. However published evaluations of ML-based AD have only had limited data and have not provided insight into why it works. To address limited evaluation against real-world data, we apply autoencoder, an existing ML-AD model, to 57 DDoS attack events captured at 5 cloud IPs from a major cloud provider. To improve our understanding for why ML-based AD works or not works, we interpret this data with feature attribution and counterfactual explanation. We show that our version of autoencoders work well overall: our models capture nearly all malicious flows to 2 of the 4 cloud IPs under attacks (at least 99.99%) but generate a few false negatives (5% and 9%) for the remaining 2 IPs. We show that our models maintain near-zero false positives on benign flows to all 5 IPs. Our interpretation of results shows that our models identify almost all malicious flows with non-whitelisted (non-WL) destination ports (99.92%) by learning the full list of benign destination ports from training data (the normality). Interpretation shows that although our models learn incomplete normality for protocols and source ports, they still identify most malicious flows with non-WL protocols and blacklisted (BL) source ports (100.0% and 97.5%) but risk false positives. Interpretation also shows that our models only detect a few malicious flows with BL packet sizes (8.5%) by incorrectly inferring these BL sizes as normal based on incomplete normality learned. We find our models still detect a quarter of flows (24.7%) with abnormal payload contents even when they do not see payload by combining anomalies from multiple flow features. Lastly, we summarize the implications of what we learn on applying autoencoder-based AD in production.problme?Machine-learning-based anomaly detection (ML-based AD) has been successful at detecting DDoS events in the lab. However published evaluations of ML-based AD have only had limited data and have not provided insight into why it works. To address limited evaluation against real-world data, we apply autoencoder, an existing ML-AD model, to 57 DDoS attack events captured at 5 cloud IPs from a major cloud provider. To improve our understanding for why ML-based AD works or not works, we interpret this data with feature attribution and counterfactual explanation. We show that our version of autoencoders work well overall: our models capture nearly all malicious flows to 2 of the 4 cloud IPs under attacks (at least 99.99%) but generate a few false negatives (5% and 9%) for the remaining 2 IPs. We show that our models maintain near-zero false positives on benign flows to all 5 IPs. Our interpretation of results shows that our models identify almost all malicious flows with non-whitelisted (non-WL) destination ports (99.92%) by learning the full list of benign destination ports from training data (the normality). Interpretation shows that although our models learn incomplete normality for protocols and source ports, they still identify most malicious flows with non-WL protocols and blacklisted (BL) source ports (100.0% and 97.5%) but risk false positives. Interpretation also shows that our models only detect a few malicious flows with BL packet sizes (8.5%) by incorrectly inferring these BL sizes as normal based on incomplete normality learned. We find our models still detect a quarter of flows (24.7%) with abnormal payload contents even when they do not see payload by combining anomalies from multiple flow features. Lastly, we summarize the implications of what we learn on applying autoencoder-based AD in production.
@TechReport{Guo19a,
author = "Hang Guo and Xun Fan and Anh Cao and Geoff
Outhred and John Heidemann",
title = "Peek Inside the Closed World: Evaluating Autoencoder-Based Detection of {DDoS} to Cloud",
institution = "arXiv",
year = 2019,
sortdate = "2019-12-16",
project = "ant, lacanic",
jsubject = "topology_modeling",
number = "arXiv:1912.05590v2 [cs.NI]",
month = dec,
jlocation = "johnh: pafile",
keywords = "ddos, cloud, machine learning, autoencoder",
url = "https://ant.isi.edu/%7ejohnh/PAPERS/Guo19a.html",
otherurl = "https://ant.isi.edu/%7ehangguo/papers/Guo19a.pdf",
pdfurl = "https://ant.isi.edu/%7ejohnh/PAPERS/Guo19a.pdf",
blogurl = "https://ant.isi.edu/blog/?p=1401",
abstract = "Machine-learning-based anomaly detection (ML-based AD) has been
successful at detecting DDoS events in the lab. However published
evaluations of ML-based AD have only had limited data and have not
provided insight into why it works. To address limited evaluation
against real-world data, we apply autoencoder, an existing ML-AD
model, to 57 DDoS attack events captured at 5 cloud IPs from a major
cloud provider. To improve our understanding for why ML-based AD works
or not works, we interpret this data with feature attribution and
counterfactual explanation. We show that our version of autoencoders
work well overall: our models capture nearly all malicious flows to 2
of the 4 cloud IPs under attacks (at least 99.99\%) but generate a few
false negatives (5\% and 9\%) for the remaining 2 IPs. We show that our
models maintain near-zero false positives on benign flows to all 5
IPs. Our interpretation of results shows that our models identify
almost all malicious flows with non-whitelisted (non-WL) destination
ports (99.92\%) by learning the full list of benign destination ports
from training data (the normality). Interpretation shows that although
our models learn incomplete normality for protocols and source ports,
they still identify most malicious flows with non-WL protocols and
blacklisted (BL) source ports (100.0\% and 97.5\%) but risk false
positives. Interpretation also shows that our models only detect a few
malicious flows with BL packet sizes (8.5\%) by incorrectly inferring
these BL sizes as normal based on incomplete normality learned. We
find our models still detect a quarter of flows (24.7\%) with abnormal
payload contents even when they do not see payload by combining
anomalies from multiple flow features. Lastly, we summarize the
implications of what we learn on applying autoencoder-based AD in
production.problme?Machine-learning-based anomaly detection (ML-based
AD) has been successful at detecting DDoS events in the lab. However
published evaluations of ML-based AD have only had limited data and
have not provided insight into why it works. To address limited
evaluation against real-world data, we apply autoencoder, an existing
ML-AD model, to 57 DDoS attack events captured at 5 cloud IPs from a
major cloud provider. To improve our understanding for why ML-based AD
works or not works, we interpret this data with feature attribution
and counterfactual explanation. We show that our version of
autoencoders work well overall: our models capture nearly all
malicious flows to 2 of the 4 cloud IPs under attacks (at least
99.99\%) but generate a few false negatives (5\% and 9\%) for the
remaining 2 IPs. We show that our models maintain near-zero false
positives on benign flows to all 5 IPs. Our interpretation of results
shows that our models identify almost all malicious flows with
non-whitelisted (non-WL) destination ports (99.92\%) by learning the
full list of benign destination ports from training data (the
normality). Interpretation shows that although our models learn
incomplete normality for protocols and source ports, they still
identify most malicious flows with non-WL protocols and blacklisted
(BL) source ports (100.0\% and 97.5\%) but risk false
positives. Interpretation also shows that our models only detect a few
malicious flows with BL packet sizes (8.5\%) by incorrectly inferring
these BL sizes as normal based on incomplete normality learned. We
find our models still detect a quarter of flows (24.7\%) with abnormal
payload contents even when they do not see payload by combining
anomalies from multiple flow features. Lastly, we summarize the
implications of what we learn on applying autoencoder-based AD in
production.",
}
Downloads: 0
{"_id":"ia54Q2mN54E44DueS","bibbaseid":"guo-fan-cao-outhred-heidemann-peekinsidetheclosedworldevaluatingautoencoderbaseddetectionofddostocloud-2019","author_short":["Guo, H.","Fan, X.","Cao, A.","Outhred, G.","Heidemann, J."],"bibdata":{"bibtype":"techreport","type":"techreport","author":[{"firstnames":["Hang"],"propositions":[],"lastnames":["Guo"],"suffixes":[]},{"firstnames":["Xun"],"propositions":[],"lastnames":["Fan"],"suffixes":[]},{"firstnames":["Anh"],"propositions":[],"lastnames":["Cao"],"suffixes":[]},{"firstnames":["Geoff"],"propositions":[],"lastnames":["Outhred"],"suffixes":[]},{"firstnames":["John"],"propositions":[],"lastnames":["Heidemann"],"suffixes":[]}],"title":"Peek Inside the Closed World: Evaluating Autoencoder-Based Detection of DDoS to Cloud","institution":"arXiv","year":"2019","sortdate":"2019-12-16","project":"ant, lacanic","jsubject":"topology_modeling","number":"arXiv:1912.05590v2 [cs.NI]","month":"December","jlocation":"johnh: pafile","keywords":"ddos, cloud, machine learning, autoencoder","url":"https://ant.isi.edu/%7ejohnh/PAPERS/Guo19a.html","otherurl":"https://ant.isi.edu/%7ehangguo/papers/Guo19a.pdf","pdfurl":"https://ant.isi.edu/%7ejohnh/PAPERS/Guo19a.pdf","blogurl":"https://ant.isi.edu/blog/?p=1401","abstract":"Machine-learning-based anomaly detection (ML-based AD) has been successful at detecting DDoS events in the lab. However published evaluations of ML-based AD have only had limited data and have not provided insight into why it works. To address limited evaluation against real-world data, we apply autoencoder, an existing ML-AD model, to 57 DDoS attack events captured at 5 cloud IPs from a major cloud provider. To improve our understanding for why ML-based AD works or not works, we interpret this data with feature attribution and counterfactual explanation. We show that our version of autoencoders work well overall: our models capture nearly all malicious flows to 2 of the 4 cloud IPs under attacks (at least 99.99%) but generate a few false negatives (5% and 9%) for the remaining 2 IPs. We show that our models maintain near-zero false positives on benign flows to all 5 IPs. Our interpretation of results shows that our models identify almost all malicious flows with non-whitelisted (non-WL) destination ports (99.92%) by learning the full list of benign destination ports from training data (the normality). Interpretation shows that although our models learn incomplete normality for protocols and source ports, they still identify most malicious flows with non-WL protocols and blacklisted (BL) source ports (100.0% and 97.5%) but risk false positives. Interpretation also shows that our models only detect a few malicious flows with BL packet sizes (8.5%) by incorrectly inferring these BL sizes as normal based on incomplete normality learned. We find our models still detect a quarter of flows (24.7%) with abnormal payload contents even when they do not see payload by combining anomalies from multiple flow features. Lastly, we summarize the implications of what we learn on applying autoencoder-based AD in production.problme?Machine-learning-based anomaly detection (ML-based AD) has been successful at detecting DDoS events in the lab. However published evaluations of ML-based AD have only had limited data and have not provided insight into why it works. To address limited evaluation against real-world data, we apply autoencoder, an existing ML-AD model, to 57 DDoS attack events captured at 5 cloud IPs from a major cloud provider. To improve our understanding for why ML-based AD works or not works, we interpret this data with feature attribution and counterfactual explanation. We show that our version of autoencoders work well overall: our models capture nearly all malicious flows to 2 of the 4 cloud IPs under attacks (at least 99.99%) but generate a few false negatives (5% and 9%) for the remaining 2 IPs. We show that our models maintain near-zero false positives on benign flows to all 5 IPs. Our interpretation of results shows that our models identify almost all malicious flows with non-whitelisted (non-WL) destination ports (99.92%) by learning the full list of benign destination ports from training data (the normality). Interpretation shows that although our models learn incomplete normality for protocols and source ports, they still identify most malicious flows with non-WL protocols and blacklisted (BL) source ports (100.0% and 97.5%) but risk false positives. Interpretation also shows that our models only detect a few malicious flows with BL packet sizes (8.5%) by incorrectly inferring these BL sizes as normal based on incomplete normality learned. We find our models still detect a quarter of flows (24.7%) with abnormal payload contents even when they do not see payload by combining anomalies from multiple flow features. Lastly, we summarize the implications of what we learn on applying autoencoder-based AD in production.","bibtex":"@TechReport{Guo19a,\n author = \"Hang Guo and Xun Fan and Anh Cao and Geoff\n Outhred and John Heidemann\",\n title = \"Peek Inside the Closed World: Evaluating Autoencoder-Based Detection of {DDoS} to Cloud\",\n\tinstitution = \t\"arXiv\",\n year = 2019,\n\tsortdate = \t\t\"2019-12-16\", \n\tproject = \"ant, lacanic\",\n\tjsubject = \"topology_modeling\",\n number = \"arXiv:1912.05590v2 [cs.NI]\",\n month = dec,\n\tjlocation = \t\"johnh: pafile\",\n\tkeywords = \t\"ddos, cloud, machine learning, autoencoder\",\n\turl =\t\t\"https://ant.isi.edu/%7ejohnh/PAPERS/Guo19a.html\",\n\totherurl =\t\t\"https://ant.isi.edu/%7ehangguo/papers/Guo19a.pdf\",\n\tpdfurl =\t\"https://ant.isi.edu/%7ejohnh/PAPERS/Guo19a.pdf\",\n\tblogurl = \"https://ant.isi.edu/blog/?p=1401\",\n abstract = \"Machine-learning-based anomaly detection (ML-based AD) has been\nsuccessful at detecting DDoS events in the lab. However published\nevaluations of ML-based AD have only had limited data and have not\nprovided insight into why it works. To address limited evaluation\nagainst real-world data, we apply autoencoder, an existing ML-AD\nmodel, to 57 DDoS attack events captured at 5 cloud IPs from a major\ncloud provider. To improve our understanding for why ML-based AD works\nor not works, we interpret this data with feature attribution and\ncounterfactual explanation. We show that our version of autoencoders\nwork well overall: our models capture nearly all malicious flows to 2\nof the 4 cloud IPs under attacks (at least 99.99\\%) but generate a few\nfalse negatives (5\\% and 9\\%) for the remaining 2 IPs. We show that our\nmodels maintain near-zero false positives on benign flows to all 5\nIPs. Our interpretation of results shows that our models identify\nalmost all malicious flows with non-whitelisted (non-WL) destination\nports (99.92\\%) by learning the full list of benign destination ports\nfrom training data (the normality). Interpretation shows that although\nour models learn incomplete normality for protocols and source ports,\nthey still identify most malicious flows with non-WL protocols and\nblacklisted (BL) source ports (100.0\\% and 97.5\\%) but risk false\npositives. Interpretation also shows that our models only detect a few\nmalicious flows with BL packet sizes (8.5\\%) by incorrectly inferring\nthese BL sizes as normal based on incomplete normality learned. We\nfind our models still detect a quarter of flows (24.7\\%) with abnormal\npayload contents even when they do not see payload by combining\nanomalies from multiple flow features. Lastly, we summarize the\nimplications of what we learn on applying autoencoder-based AD in\nproduction.problme?Machine-learning-based anomaly detection (ML-based\nAD) has been successful at detecting DDoS events in the lab. However\npublished evaluations of ML-based AD have only had limited data and\nhave not provided insight into why it works. To address limited\nevaluation against real-world data, we apply autoencoder, an existing\nML-AD model, to 57 DDoS attack events captured at 5 cloud IPs from a\nmajor cloud provider. To improve our understanding for why ML-based AD\nworks or not works, we interpret this data with feature attribution\nand counterfactual explanation. We show that our version of\nautoencoders work well overall: our models capture nearly all\nmalicious flows to 2 of the 4 cloud IPs under attacks (at least\n99.99\\%) but generate a few false negatives (5\\% and 9\\%) for the\nremaining 2 IPs. We show that our models maintain near-zero false\npositives on benign flows to all 5 IPs. Our interpretation of results\nshows that our models identify almost all malicious flows with\nnon-whitelisted (non-WL) destination ports (99.92\\%) by learning the\nfull list of benign destination ports from training data (the\nnormality). Interpretation shows that although our models learn\nincomplete normality for protocols and source ports, they still\nidentify most malicious flows with non-WL protocols and blacklisted\n(BL) source ports (100.0\\% and 97.5\\%) but risk false\npositives. Interpretation also shows that our models only detect a few\nmalicious flows with BL packet sizes (8.5\\%) by incorrectly inferring\nthese BL sizes as normal based on incomplete normality learned. We\nfind our models still detect a quarter of flows (24.7\\%) with abnormal\npayload contents even when they do not see payload by combining\nanomalies from multiple flow features. Lastly, we summarize the\nimplications of what we learn on applying autoencoder-based AD in\nproduction.\",\n}\n\n","author_short":["Guo, H.","Fan, X.","Cao, A.","Outhred, G.","Heidemann, J."],"bibbaseid":"guo-fan-cao-outhred-heidemann-peekinsidetheclosedworldevaluatingautoencoderbaseddetectionofddostocloud-2019","role":"author","urls":{"Paper":"https://ant.isi.edu/%7ejohnh/PAPERS/Guo19a.html"},"keyword":["ddos","cloud","machine learning","autoencoder"],"metadata":{"authorlinks":{}}},"bibtype":"techreport","biburl":"https://bibbase.org/f/dHevizJoWEhWowz8q/johnh-2023-2.bib","dataSources":["YLyu3mj3xsBeoqiHK","fLZcDgNSoSuatv6aX","fxEParwu2ZfurScPY","7nuQvtHTqKrLmgu99"],"keywords":["ddos","cloud","machine learning","autoencoder"],"search_terms":["peek","inside","closed","world","evaluating","autoencoder","based","detection","ddos","cloud","guo","fan","cao","outhred","heidemann"],"title":"Peek Inside the Closed World: Evaluating Autoencoder-Based Detection of DDoS to Cloud","year":2019}