A secured database monitoring method to improve data backup and recovery operations in cloud computing

G. Ramesh1, J. Logeshwaran2* and V. Aravindarajan1

*Correspondence:
J. Logeshwaran,
drlogeshj@gmail.com

Received: 20 March 2023; Accepted: 15 April 2023; Published: 24 April 2023.

In general, the company sometimes uses unregistered functions in database, which significantly improves performance, but does not leave the possibility of recovery except for backup. That is, actions must be performed immediately after passing the session. A queue problem is likely to cause data loss and downtime of about a week. In modern conditions, this can lead to the bankruptcy of the company. It can be seen that backup systems have been installed and configured, but despite this, they have not succeeded in restoring within the time frame specified in the SLA. In this study, a secured database monitoring method was proposed to improve data backup and recovery operations in cloud computing. In this proposed method, the backup speed is directly proportional to the amount of data, while having at least 30% annual data growth. In 3–4 years, the data at least doubled, but for some companies, this number is even higher, while the backup speed does not change. Those terms and those SLAs that were relevant 3–4 years ago now need to be at least doubled. At the same time, business requirements for data recovery (recovery point objective/recovery time objective) continue to grow.

Keywords: database, performance, recovery, backup, operating system, queue, downtime

Introduction

All business processes of the company are being transferred to information technology, and the paper primary (copies and originals of documents, reports, scans, etc.) is disappearing (1, 2). Everything revolves around IT systems, and data loss literally means losing everything (3). There is no more room for error. In the cases we presented, all the time, companies were unable to function due to the unavailability of data due to various circumstances (4, 5). This led to direct and indirect losses, for example, reputational losses, when the company could not carry out its main business process, which is not easy to measure in monetary terms but in the long run does not cause less damage to the company (68). We have reflected on our observations regarding recovery time objective. As your data grow, the actual recovery time must also grow, and the SLA requirements become stricter (9). The point where the actual graph time equals the required time has already been sent to most clients (10). In fact, most errors are associated with the loss of some part of the data. At the same time, traditional backup tools allow you to restore data directly from the backup, but often you need to restore the entire system (1113). If your database is 15 TB, this will take several days (14). In a world where the master sheet disappears and everything is stored in IT systems, every second creates data that we want to protect immediately, the moment it is created. But this cannot be done using classic backup systems (15, 16). Each data have a certain length of time, and these data exist all over the world in one copy (17).

The customers want to continue to protect their data from the moment it appears (18). When you decide to restore from a backup, you often need to restore a day earlier and then have the data retrieved somewhere else a day later (19, 20). As a rule, this is a long task for administrators that take several days (21). With a very negative development of events, it can lead to the loss of very important information (22). Of course, the problem is not limited to backup; it is about the construction of the IT system as a whole, and the title in this case is so important that it cannot be ignored (23, 24). Unfortunately, there are no cheap and quick options to check how well the backup was done (25). Of course, this can be done with periodic test recoveries, but this is a very expensive operation in terms of human effort and IT resources (26). It is the work of a separate team on a separate hardware (27). Most of the customers do not do this. It often happens that everyone creates backups, but during restoration, it turns out that they could not have been done – despite the outwardly correct operation of RMS, they cannot be restored (28, 29). This happens for various reasons. This is best explained with an example (30). A backup system is a service subsystem of a data center and has the following features:

• The backup process is not critical to solving IS problems. A failure of a backup system does not lead to a reduction in the availability of critical information services (31).

• The computational burden that the backup process creates is not beneficial to the provision of IS information services (32).

Naturally, it was forgotten in the maintenance industry for many years. Thus, a part of the data is completely lost (33, 34). It took several more days to completely recreate the infrastructure services – no backups of operating systems, binaries, configurations, etc. (35). All the missing information was collected from primary documents (36). One of our customers used a SAP system with an Oracle database. The backup was carried out by inbuilt SAP tools with the help of huge vendors (37, 38). Two different backup policies were configured: the first was file-based and copied the data of operating and software systems, and the second was the database itself (39). Since they were sent to the same system, an exclusion list was constructed and entered into the database (40). The file policy takes this list into account and does not allocate directories in the database (41). Due to the uniqueness of the architecture, the database backup policy ignored the list of exclusions and correctly copied the required data (42).

Related works

An improper approach to the problem of backup is a critical problem. Historically, the company or an integrator built it. At the time of construction, it certainly met all the requirements and performed its function perfectly (43). Over the years, the enterprise IT landscape has changed. At the same time, the backup system was simply adjusted as the system evolved, and often no systematic approach was followed, which would take into account the importance of system compliance with initial indicators in all subsequent phases (44). When developing a data protection system in your organization, you should take into account that this is only one part of your data protection strategy (45). A reserve parachute is not a silver bullet, as many case studies prove that the approach to data protection must be comprehensive, so when developing it, we must consider where it will fit in the global strategy for data protection (46).

A system administrator maintains an inventory of backup client computers, recording devices, and backup storage media and schedules backups. All this information is contained in a special database stored on the backup management server (47). The management server instructs the agent program installed on the client computer to start backing up data according to the selected policy in accordance with the schedule or according to the operator’s command (48). The agent program collects and transfers the data to be backed up to the copy server specified by the management server (49). The mechanism of action of “snapshots” is different and can be implemented both in software on a production server and in hardware within the array (50). At the moment when the backup needs to be started, the agent program instructs the application to complete all transactions and save the cache memory (51). A fundamental challenge for any data center is implementing a service level agreement between the IT department and the business (52). A key aspect of meeting business needs is the assurance of data security, so a data backup and recovery system is an integrated infrastructure unit of a properly organized data center’s data storage subsystem (53). This type of backup and recovery actually takes a few seconds, which distinguishes this technology from classical systems with foreign media (54). The issues in the existing systems and the proposed solutions are shown in Table 1.

TABLE 1
www.bohrpub.com

Table 1. The related issues and proposed suggestions.

Proposed model

First, it is necessary to disconnect the backup and recovery speed from the computer volume. Data storage systems, application software, and RMS manufacturers recommend using some tools that can be used to solve this problem. Snapshots allow you to back up and restore data in seconds with little or no impact on performance. This is done through sequence, while being able to control SRK, and should be part of its policy. Another solution may be to use various utility tools like Oracle Standby, DB2 HADR, and MS SQL Always On. All of these tools allow you to have a working copy of the production system, separated from the original, which can be deployed immediately. It allows you to work immediately after failures. The second is to make it possible to restore only the data you need. Our approach takes into account that when restoring a part of the data, there is no need to copy the entire system, we can restore the data we need at this time. This is achieved by the ability to quickly deploy or use existing systems that contain this data. As in the first case, the snapshot allows you to solve this problem (you can quickly open a snapshot on a neighboring server and extract the necessary data). It also includes continuous data protection technologies, for example, Oracle Standby with Flashback. They allow you to quickly deploy a working copy of your data in a timely manner.

If you need to get a logical volume, for example, a row or a database table, these tools greatly simplify the task, allowing you to restore the necessary data without restoring the entire copy. The third is to reduce the gap between the origin of data and its protection. This can be achieved in several ways, depending on the specifics of a particular case and the degree of importance of the data. For example, for less critical systems, the time interval for backups can be reduced to a few hours. In this case, we use snapshots. They can act as a restore point that can be done once per hour. Some modern arrays handle these processes very well and can store large numbers of system snapshots. This is a great way out of a situation where you need to step back for a while. For more critical systems, there may be no time interval, and data must be protected continuously. There are many solutions of this class, for example, Oracle Standby with Flash Back, which allows you to roll back the database for a while by recording all changes. You can use Oracle ZDLRA PAK, which will receive almost synchronously all changes in database, or general purpose software and hardware systems. They record all changes and allow you to restore to any point in time. Regular backups can be an indispensable tool in the event of major disasters or the need to recover data from long ago. However, in the present situation, it is only a reserve parachute, used at the last moment. The proposed model has the following important blocks (Figure 1):

FIGURE 1
www.bohrpub.com

Figure 1. Important blocks of the proposed model.

• Backup

• Recovery

• Content analysis

• Contextual search

• Mobile data access

• Seamless integration with the cloud

• Information security tasks

The fourth is to reduce hidden errors. There is only one way to make sure that the backup is working correctly – try to restore it. This is the most perfect and rarely used method by our clients. But we offer a way out of this situation. First, have easily recoverable instances of computers. This is again a story about snapshot and standby systems that can be quickly deployed and tested. It takes incomparably less time and effort than “unpacking” an entire backup. Of course, this does not always help; but in case of an emergency, at least it gives a little more confidence that the data can be recovered by these means. Second, some SRCs allow automated testing. At a specific point in time on a schedule, you can run virtual machines in an isolated environment and use predefined algorithms to verify that data have actually been retrieved, that the application is available, that it is consistent, and that it is responding to the necessary requests. In this way, administrators can be relieved of long, routine tasks. A centralized backup system has a multi-layered architecture, which includes the following, as shown in Figure 2:

FIGURE 2
www.bohrpub.com

Figure 2. Components of centralized backup system.

• A backup management server that can also incorporate the functions of a data copy server.

• One or more data copy servers to which backup devices are attached.

• Client computers that have backup agent programs installed.

• Backup System Administrator Console.

The fifth was the backup system’s transparency. The described integrated approach involves creating a complex system using different technologies from different manufacturers. The task of making this system actually work, with further changes and scaling possible, is not trivial and can be tackled in two ways:

• The first way is that the customer is self-sufficient and wants to implement the system. Here, as a coordinator, we help create all the necessary processes, create regulatory frameworks, create all the necessary instructions and plans so that the customer’s IT department can develop, and run the system in the right direction more independently. Then, transfer all this practical basis of terms and tasks to the customer in the form of functional organization of business processes.

• The second option is when the customer is not sure that the SRK system can be continuously maintained in a combat condition; the solution is to partially or fully outsource the system. We have clients who successfully use this service as both SLA requirements and our engagement levels as an IT outsourcer continue to increase.

Unfortunately, there is still no universal recipe for solving the problem of data recovery under the current conditions of constant system development and problems. Only a combination of the above solutions and a systematic approach will allow organizations to recover data at the point of business need. Backup systems ensure the continuity of business processes.

Results and discussion

The proposed database monitoring method (DMM) was compared with the existing storage backup and recovery strategy (SBRS), Distributed Data Backup and Recovery Method (DDBRM), Automation of Disaster Recovery and Security (ADRS), and control and communication management (CCM)

Data backup management

Events may have undesirable consequences for the IT infrastructure and business in general; for example, a fire in a building, the breakthrough of the central heating battery in the server room, or casual theft of equipment and components. One of the ways to avoid data loss in such situations is to store backups in a location away from the central area of the server equipment. The comparison of data backup management is shown in Figure 3.

FIGURE 3
www.bohrpub.com

Figure 3. Comparison of data backup management.

Off-server copy management

Providing a quick way to access the data needed for recovery is essential. The method described is off-site (in other words, keeping copies off-site). Two methods are used to organize this process. It has been writing data to removable media and physically moving it. In this case, you need to take care of the means to provide the press quickly in case of failure. The comparison of off-server copy management is shown in Figure 4.

FIGURE 4
www.bohrpub.com

Figure 4. Comparison of off-server copy management.

Storage device management

The advantage of this method is the ability to organize without difficulty. The downside is the difficulty of retrieving the media, the need to transfer the information to storage, and the risk of damaging the media in transit. They are copying data to another location over a network channel. The comparison of storage device management is shown in Figure 5.

FIGURE 5
www.bohrpub.com

Figure 5. Comparison of storage device management.

Server-less copy management

This method uses a VPN tunnel over the Internet. The advantage is that there is no need to transport the information somewhere; the disadvantage is the need to use a sufficiently wide channel (as a rule, it is costly) and protect the transmitted data (e.g., using the same VPN). Compression algorithms or deductive technology can significantly reduce the difficulty in converting large amounts of data. The comparison of server-less copy management is shown in Figure 6.

FIGURE 6
www.bohrpub.com

Figure 6. Comparison of server-less copy management.

Data copy management

Separately, security measures should be taken when organizing data storage. First, it is necessary to ensure that the data carriers are in a protected area and that measures are in place to prevent unauthorized persons from reading the data, for example, by using an encryption system, entering a non-disclosure agreement, and so on. The comparison of data copy management is shown in Figure 7.

FIGURE 7
www.bohrpub.com

Figure 7. Comparison of data copy management.

If removable media is involved, the data on it must also be encrypted. In this case, the marking system should not help the attacker analyze the data. It is necessary to use a faceless numbering scheme to indicate the carriers of the names of the changed files.

Conclusion

To automate backup processes, special software is used that controls the process of creating backups and the recovery process and allows you to work with various data carriers, including tape devices. What data should be stored, where, and on which bandwidth? Additional features of centralized backup software allow you to restore individual characters and database tables without restoring the entire volume of data. Writing backups to tape allows you to organize remote storage of backups and the protection of critical data in the event of a disaster. Using tape media to store archival copies allows data to be read 50 years after it was written. It will only be effective if done regularly. If data change frequently, replication should also be frequent. Therefore, rather than copying data manually, many prefer to use specialized software to automate this process. Any computer has many risks that can lead to hardware failure and loss of important data. In the business sector, this is even worse because the loss of business data results in huge losses and lost customers. When choosing a backup program, it wants to make sure that you can easily restore the data it wants.

References

1. Adhikari N, Logeshwaran J, Kiruthiga T. The artificially intelligent switching framework for terminal access provides smart routing in modern computer networks. BOHR Int J Smart Comput Inf Technol. 3:45–50. doi: 10.54646/BIJSCIT.18

CrossRef Full Text | Google Scholar

2. Zhang Y, Xu C, Muntean GM. A novel distributed data backup and recovery method for software defined-WAN controllers. Proceedings of the 2021 IEEE Global Communications Conference (GLOBECOM). Madrid: IEEE (2021). p. 01–06. doi: 10.1109/GLOBECOM46510.2021.9685291

CrossRef Full Text | Google Scholar

3. Gopi B, Logeshwaran J, Gowri J, Kiruthiga T. The moment probability and impacts monitoring for electron cloud behavior of electronic computers by using quantum deep learning model. Neuroquantology. (2022) 20:6088–100.

Google Scholar

4. Wang Q. Cloud data backup and recovery method based on the DELTA compression algorithm. Proceedings of the 2021 IEEE International Conference on Industrial Application of Artificial Intelligence (IAAI). Harbin: IEEE (2021). p. 183–8. doi: 10.1109/IAAI54625.2021.9699943

CrossRef Full Text | Google Scholar

5. Logeshwaran J. The topology configuration of protocol-based local networks in high speed communication networks. Multidisciplinary Approach Res. (2022) 15:78–83.

Google Scholar

6. Yao L. Feasibility study report on cloud technology of high-performance self-service database. In: Hung JC, Chang J, Pei Y, Wu W editors. Innovative Computing. Lecture Notes in Electrical Engineering. Singapore: Springer (2022). p. 1315–23. doi: 10.1007/978-981-16-4258-6_160

CrossRef Full Text | Google Scholar

7. Ramesh G, Logeshwaran J, Aravindarajan V. The performance evolution of antivirus security systems in ultra dense cloud server using intelligent deep learning. BOHR Int J Comput Intellig Commun Netw. (2022) 1:15–9. doi: 10.54646/bijcicn.004

CrossRef Full Text | Google Scholar

8. Jin D, Wang Q. CDP backup and recovery method for ensuring database consistency. Proceedings of the 2021 IEEE International Conference on Power Electronics, Computer Applications (ICPECA). Shenyang: IEEE (2021). p. 722–8. doi: 10.1109/ICPECA51329.2021.9362541

CrossRef Full Text | Google Scholar

9. Ramesh G, Logeshwaran J, Gowri J, Ajay Mathew. The management and reduction of digital noise in video image processing by using transmission based noise elimination scheme. ICTACT J Image Video Process. (2022) 13:2797–801.

Google Scholar

10. Adee R, Mouratidis H. A dynamic four-step data security model for data in cloud computing based on cryptography and steganography. Sensors. (2022) 22:1109. doi: 10.3390/s22031109

CrossRef Full Text | Google Scholar

11. Jasmine J, Yuvaraj N, Logeshwaran J. DSQLR-A distributed scheduling and QoS localized routing scheme for wireless sensor network. Recent Trends Inf Technol Commun Industry 4.0. (2022) 1:47–60.

Google Scholar

12. Jing F, Huang D, Wang Y, Lv X, Du Q. Remote sensing data backup and recovery system for power industry. Proceedings of the 2021 IEEE Intl Conf on Parallel & Distributed Processing with Applications, Big Data & Cloud Computing, Sustainable Computing & Communications, Social Computing & Networking (ISPA/BDCloud/SocialCom/SustainCom). New York, NY: IEEE (2021). p. 1717–23. doi: 10.1109/ISPA-BDCloud-SocialCom-SustainCom52081.2021.00231

CrossRef Full Text | Google Scholar

13. Logeshwaran J, Adhikari N, Joshi SS, Saxena P, Sharma A. The deep DNA machine learning model to classify the tumor genome of patients with tumor sequencing. Int J Health Sci. (2022) 6:9364–75. doi: 10.53730/ijhs.v6nS5.10767

CrossRef Full Text | Google Scholar

14. Ravimaran S, Sathish A. Robust security with strong authentication in mobile cloud computing based on trefoil congruity framework. J Organ End User Comput. (2021) 33:1–28. doi: 10.4018/JOEUC.20211101.oa11

CrossRef Full Text | Google Scholar

15. Logeshwaran J, Ramkumar M, Kiruthiga T, Sharan Pravin R. SVPA - the segmentation based visual processing algorithm (SVPA) for illustration enhancements in digital video processing (DVP). ICTACT J Image Video Process. (2022) 12:2669–73.

Google Scholar

16. Smara M, Aliouat M, Harous S, Pathan ASK. Robustness improvement of component-based cloud computing systems. J Super Comput. (2022) 78:4977–5009. doi: 10.1007/s11227-021-04054-2

CrossRef Full Text | Google Scholar

17. Ramesh G, Aravindarajan V, Logeshwaran J, Kiruthiga T, Vignesh S. Estimation analysis of paralysis effects for human nervous system by using Neuro fuzzy logic controller. Neuroquantology. (2022) 20:3195–206.

Google Scholar

18. Ampera D, Fibriasari H. Implementation of cloud computing system in learning system development in engineering education study program. Int J Educ Math Sci Technol. (2021) 9:728–40. doi: 10.46328/ijemst.2114

CrossRef Full Text | Google Scholar

19. Logeshwaran J, Ramkumar M, Kiruthiga T, Sharanpravin R. The role of integrated structured cabling system (ISCS) for reliable bandwidth optimization in high-speed communication network. ICTACT J Commun Technol. (2022) 13:2635–9.

Google Scholar

20. Jeslin JG, Kumar PM. Implementing an efficient data deduplication framework for cloud storage. Indian J. Comput. Sci. Eng. (2022) 13:136–44. doi: 10.21817/indjcse/2022/v13i1/221301046

CrossRef Full Text | Google Scholar

21. Logeshwaran J, Rex MJ, Kiruthiga T, Rajan VA. FPSMM: Fuzzy probabilistic based semi morkov model among the sensor nodes for realtime applications. Proceedings of the 2017 International Conference on Intelligent Sustainable Systems (ICISS). Palladam: IEEE (2017). p. 442–6. doi: 10.1109/ISS1.2017.8389391

CrossRef Full Text | Google Scholar

22. Zhao J, Zhang L, Zhao Y. Informatization of accounting systems in small-and medium-sized enterprises based on artificial intelligence-enabled cloud computing. Comput Intellig Neurosci. (2022) 2022. doi: 10.1155/2022/6089195

CrossRef Full Text | Google Scholar

23. Raja S, Logeshwaran J, Venkatasubramanian S, Jayalakshmi M, Rajeswari N, Olaiya NG, et al. OCHSA: Designing energy-efficient lifetime-aware leisure degree adaptive routing protocol with optimal cluster head selection for 5G communication network disaster management. Sci. Program. (2022) 2022. doi: 10.1155/2022/5424356

CrossRef Full Text | Google Scholar

24. Shaik V, Natarajan K. Asynchronous method of oracle: A cost-effective and reliable model for cloud migration using incremental backups. Data Science and Security. Singapore: Springer (2021). p. 417–27. doi: 10.1007/978-981-16-4486-3_47

CrossRef Full Text | Google Scholar

25. Ramesh G, Logeshwaran J, Rajkumar K. The smart construction for image preprocessing of mobile robotic systems using neuro fuzzy logical system approach. Neuroquantology. (2022) 20:6354–67.

Google Scholar

26. Belgacem A. Dynamic resource allocation in cloud computing: Analysis and taxonomies. Computing. (2022) 104:681–710. doi: 10.1007/s00607-021-01045-2

CrossRef Full Text | Google Scholar

27. Gopi B, Ramesh G, Logeshwaran J. An innovation for energy release of nuclear fusion at short distance dielectrics in semiconductor model. ICTACT J Microelectron. (2022) 8:1430–5.

Google Scholar

28. Zhou J, Zou W. Construction and application of Wenhua education cloud based on Huawei FusionSphere. Proceedings of the 2021 IEEE International Conference on Artificial Intelligence and Computer Applications (ICAICA). Dalian: IEEE (2021). p. 477–81. doi: 10.1109/ICAICA52286.2021.9498210

CrossRef Full Text | Google Scholar

29. Logeshwaran J. AICSA - an artificial intelligence cyber security algorithm for cooperative P2P file sharing in social networks. ICTACT J Data Sci Mach Learn. (2021) 3:251–3.

Google Scholar

30. Chang D, Li L, Chang Y, Qiao Z. Cloud computing storage backup and recovery strategy based on secure IoT and spark. Mobile Inf Syst. (2021) 2021:9505249. doi: 10.1155/2021/9505249

CrossRef Full Text | Google Scholar

31. Logeshwaran J, Malik JA, Adhikari N, Joshi SS, Bishnoi P. IoT-TPMS: An innovation development of triangular patient monitoring system using medical internet of things. Int J Health Sci. (2022) 6:9070–84. doi: 10.53730/ijhs.v6nS5.10765

CrossRef Full Text | Google Scholar

32. Wang F, Wang H, Xue L. Research on data security in big data cloud computing environment. Proceedings of the 2021 IEEE 5th Advanced Information Technology, Electronic and Automation Control Conference (IAEAC). (Vol. 5), Chongqing: IEEE (2021). p. 1446–50. doi: 10.1109/IAEAC50856.2021.9391025

CrossRef Full Text | Google Scholar

33. Gopi B, Ramesh G, Logeshwaran J. The fuzzy logical controller based energy storage and conservation model to achieve maximum energy efficiency in modern 5g communication. ICTACT J Commun Technol. (2022) 13:2774–9.

Google Scholar

34. Tiwari PK, Pandey SK, Thamba Meshach W, Parashar J, Kumar A, Altuwairiqi M, et al. Improved data security in cloud environment for test automation framework and access control for industry 4.0. Wireless Commun Mobile Comput. (2022) 2022. doi: 10.1155/2022/3242092

CrossRef Full Text | Google Scholar

35. Sekar G, Sivakumar C, Logeshwaran J. NMLA: The smart detection of motor neuron disease and analyze the health impacts with neuro machine learning model. Neuroquantology. (2022) 20:892–9.

Google Scholar

36. Peng X, Song Y, Yuan K, Guo X, Lu Z, Liu J. UCloudStack-A private cloud platform for lightweight delivery. Proceedings of the 2022 IEEE 7th International Conference on Smart Cloud (SmartCloud). Shanghai: IEEE (2022). p. 184–9. doi: 10.1109/SmartCloud55982.2022.00036

CrossRef Full Text | Google Scholar

37. Ramkumar M, Logeshwaran J, Husna T. CEA: Certification based encryption algorithm for enhanced data protection in social networks. Fundamentals Appl Math Soft Comput. (2022) 1:161–70.

Google Scholar

38. Perri D, Simonetti M, Tasso S, Ragni F, Gervasi O. Implementing a scalable and elastic computing environment based on cloud containers. Proceedings of the International Conference on Computational Science and Its Applications. Cham: Springer (2021). p. 676–89. doi: 10.1007/978-3-030-86653-2_49

CrossRef Full Text | Google Scholar

39. Logeshwaran J, Karthick S. A smart design of a multi-dimensional antenna to enhance the maximum signal clutch to the allowable standards in 5G communication networks. ICTACT J Microelectron. (2022) 8:1269–74.

Google Scholar

40. Shareef TH, Sharif KH, Rashid BN. A survey of comparison different cloud database performance: SQL and NoSQL. Passer J Basic Appl Sci. (2022) 4:45–57. doi: 10.24271/psr.2022.301247.1104

CrossRef Full Text | Google Scholar

41. Logeshwaran J. The control and communication management for ultra dense cloud system using fast Fourier algorithm. ICTACT J Data Sci Mach Learn. (2022) 3:281–4.

Google Scholar

42. Subramani R, Kaliappan S, Arul PV, Sekar S, Poures MV, De Patil PP, et al. A recent trend on additive manufacturing sustainability with supply chain management concept, multicriteria decision making techniques. Adv Mater Sci Eng. (2022) 2022:9151839. doi: 10.1155/2022/9151839

CrossRef Full Text | Google Scholar

43. Logeshwaran J, Shanmugasundaram RN. Enhancements of resource management for device to device (D2D) communication: A review. Proceedings of the 2019 Third International conference on I-SMAC (IoT in Social, Mobile, Analytics and Cloud)(I-SMAC). Palladam: IEEE (2019). p. 51–5. doi: 10.1109/I-SMAC47947.2019.9032632

CrossRef Full Text | Google Scholar

44. Raja S, Rajan AJ. A decision-making model for selection of the suitable FDM machine using fuzzy TOPSIS. Math Probl Eng. (2022) 2022:7653292. doi: 10.1155/2022/7653292

CrossRef Full Text | Google Scholar

45. Ramesh G, Logeshwaran J, Aravindarajan V, Feny Thachil. Eliminate the interference in 5g ultra-wide band communication antennas in cloud computing networks. ICTACT J Microelectron. (2022) 8:1338–44.

Google Scholar

46. Raja S, Rajan AJ, Kumar VP, Rajeswari N, Girija M, Modak S, et al. Selection of additive manufacturing machine using analytical hierarchy process. Sci. Program. (2022) 2022:1596590. doi: 10.1155/2022/1596590

CrossRef Full Text | Google Scholar

47. Gopi B, Logeshwaran J, Gowri J, Aravindarajan V. The Identification of quantum effects in electronic devices based on charge transfer magnetic field model. Neuroquantology. (2022) 20:5999–6010.

Google Scholar

48. Venkatasubramanian S, Raja S, Sumanth V, Dwivedi JN, Sathiaparkavi J, Modak S, et al. Fault diagnosis using data fusion with ensemble deep learning technique in IIoT. Math Probl Eng. (2022) 2022:1682874. doi: 10.1155/2022/1682874

CrossRef Full Text | Google Scholar

49. Vaniprabha A, Logeshwaran J, Kiruthiga T, Shah KB. Examination of the effects of long-term COVID-19 impacts on patients with neurological disabilities using a neuro machine learning model. BOHR Int J Neurol Neurosci. (2022) 1:21–8. doi: 10.54646/bijnn.003

CrossRef Full Text | Google Scholar

50. Nasurudeen TFK, Shukla VK, Gupta S. Automation of disaster recovery and security in cloud computing. Proceedings of the 2021 International Conference on Communication information and Computing Technology (ICCICT). Mumbai: IEEE (2021). p. 1–6.

Google Scholar

51. Gopi B, Logeshwaran J, Kiruthiga T. An innovation in the development of a mobile radio model for a dual-band transceiver in wireless cellular communication. BOHR Int J Comput Intellig Commun Netw. (2022) 1:20–5. doi: 10.54646/bijcicn.005

CrossRef Full Text | Google Scholar

52. Saravanakumar K, Logeshwaran J. Auto-theft prevention system for underwater sensor using lab view. Int J Innov Res Comput Commun Eng. (2016) 4:1750–5.

Google Scholar

53. Sutharasan M, Logeshwaran J. Design intelligence data gathering and incident response model for data security using honey pot system. Int J Res Dev Technol. (2016) 5:310–4.

Google Scholar