epoll.c 19 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686687688689690691692693694695696697698699700701702703704705706707708709710711712713714715716717718719720721722723724725726727728729730731732733734735736737738739740741742743744745746747748749750751752753754755756757758759760761762763764765766767768769770771772773774775776777778779780781782783784785786787788789790791792793794795796797798799800801802803804805806807808809
  1. /*
  2. * Copyright (c) 2006-2023, RT-Thread Development Team
  3. *
  4. * SPDX-License-Identifier: Apache-2.0
  5. *
  6. * Change Logs:
  7. * Date Author Notes
  8. * 2023-07-29 zmq810150896 first version
  9. */
  10. #include <rtthread.h>
  11. #include <fcntl.h>
  12. #include <stdint.h>
  13. #include <unistd.h>
  14. #include <dfs_file.h>
  15. #include "sys/epoll.h"
  16. #include "poll.h"
  17. #include <lwp_signal.h>
  18. #define EPOLL_MUTEX_NAME "EVENTEPOLL"
  19. #define EFD_SHARED_EPOLL_TYPE (EPOLL_CTL_ADD | EPOLL_CTL_DEL | EPOLL_CTL_MOD)
  20. #define EPOLLINOUT_BITS (EPOLLIN | EPOLLOUT | EPOLLRDNORM | EPOLLWRNORM)
  21. #define EPOLLEXCLUSIVE_BITS (EPOLLINOUT_BITS | EPOLLERR | EPOLLHUP | \
  22. EPOLLET | EPOLLEXCLUSIVE)
  23. static struct rt_spinlock spinlock;
  24. struct rt_eventpoll;
  25. /* Monitor queue */
  26. struct rt_fd_list
  27. {
  28. rt_uint32_t revents; /* Monitored events */
  29. struct epoll_event epev;
  30. rt_pollreq_t req;
  31. struct rt_eventpoll *ep;
  32. struct rt_wqueue_node wqn;
  33. struct rt_fd_list *next;
  34. };
  35. struct rt_ready_list
  36. {
  37. int exclusive;/* If triggered horizontally, a check is made to see if the data has been read, and if there is any data left to read, the readability event is returned in the next epoll_wait */
  38. struct rt_fd_list *rdl_event; /* rdl: ready list */
  39. struct rt_ready_list *next;
  40. };
  41. struct rt_eventpoll
  42. {
  43. rt_uint32_t tirggered; /* the waited thread whether triggered */
  44. rt_wqueue_t epoll_read;
  45. rt_thread_t polling_thread;
  46. struct rt_mutex lock;
  47. struct rt_fd_list *fdlist; /* Monitor list */
  48. int eventpoll_num; /* Number of ready lists */
  49. rt_pollreq_t req;
  50. struct rt_ready_list *rdlist; /* ready list */
  51. };
  52. static int epoll_close(struct dfs_file *file);
  53. static int epoll_poll(struct dfs_file *file, struct rt_pollreq *req);
  54. static int epoll_get_event(struct rt_fd_list *fl, rt_pollreq_t *req);
  55. static int epoll_do_ctl(int epfd, int op, int fd, struct epoll_event *event);
  56. static const struct dfs_file_ops epoll_fops =
  57. {
  58. .close = epoll_close,
  59. .poll = epoll_poll,
  60. };
  61. static int epoll_close_fdlist(struct rt_fd_list *fdlist)
  62. {
  63. struct rt_fd_list *fre_node, *list;
  64. if (fdlist != RT_NULL)
  65. {
  66. list = fdlist;
  67. while (list->next != RT_NULL)
  68. {
  69. fre_node = list->next;
  70. rt_wqueue_remove(&fre_node->wqn);
  71. list->next = fre_node->next;
  72. rt_free(fre_node);
  73. }
  74. rt_free(fdlist);
  75. }
  76. return 0;
  77. }
  78. static int epoll_close_rdlist(struct rt_ready_list *rdlist)
  79. {
  80. struct rt_ready_list *list, *fre_node;
  81. list = rdlist;
  82. if (list)
  83. {
  84. while (list->next != RT_NULL)
  85. {
  86. fre_node = list->next;
  87. list->next = fre_node->next;
  88. rt_free(fre_node);
  89. }
  90. rt_free(rdlist);
  91. }
  92. return 0;
  93. }
  94. static int epoll_close(struct dfs_file *file)
  95. {
  96. struct rt_eventpoll *ep;
  97. if (file->vnode)
  98. {
  99. if (file->vnode->data)
  100. {
  101. ep = file->vnode->data;
  102. if (ep)
  103. {
  104. rt_mutex_take(&ep->lock, RT_WAITING_FOREVER);
  105. if (ep->fdlist)
  106. {
  107. epoll_close_fdlist(ep->fdlist);
  108. }
  109. if (ep->rdlist)
  110. {
  111. epoll_close_rdlist(ep->rdlist);
  112. }
  113. rt_mutex_release(&ep->lock);
  114. rt_mutex_detach(&ep->lock);
  115. rt_free(ep);
  116. }
  117. }
  118. }
  119. return 0;
  120. }
  121. static int epoll_poll(struct dfs_file *file, struct rt_pollreq *req)
  122. {
  123. struct rt_eventpoll *ep;
  124. struct rt_fd_list *fdlist;
  125. int mask;
  126. int events = 0;
  127. if (file->vnode->data)
  128. {
  129. ep = file->vnode->data;
  130. ep->req._key = req->_key;
  131. rt_poll_add(&ep->epoll_read, req);
  132. fdlist = ep->fdlist;
  133. if (fdlist)
  134. {
  135. while (fdlist->next != RT_NULL)
  136. {
  137. fdlist = fdlist->next;
  138. mask = epoll_get_event(fdlist, &fdlist->req);
  139. if (mask & fdlist->revents)
  140. {
  141. events |= mask | POLLIN | EPOLLRDNORM;
  142. break;
  143. }
  144. }
  145. }
  146. }
  147. return events;
  148. }
  149. static int epoll_rdlist_add(struct rt_fd_list *fdl, rt_uint32_t revents)
  150. {
  151. struct rt_ready_list *rdlist = RT_NULL;
  152. struct rt_eventpoll *ep;
  153. int isexist = 0;
  154. int res = -1;
  155. ep = fdl->ep;
  156. if (revents & ep->req._key)
  157. {
  158. rt_wqueue_wakeup(&ep->epoll_read, (void*)POLLIN);
  159. }
  160. rt_mutex_take(&ep->lock, RT_WAITING_FOREVER);
  161. if (ep->rdlist == RT_NULL)
  162. {
  163. ep->rdlist = (struct rt_ready_list *)rt_malloc(sizeof(struct rt_ready_list));
  164. if (ep->rdlist == RT_NULL)
  165. {
  166. return -1;
  167. }
  168. ep->rdlist->next = RT_NULL;
  169. }
  170. rdlist = ep->rdlist;
  171. while (rdlist->next != RT_NULL)
  172. {
  173. rdlist = rdlist->next;
  174. if (rdlist->rdl_event->epev.data.fd == fdl->epev.data.fd)
  175. {
  176. isexist = 1;
  177. res = 0;
  178. break;
  179. }
  180. }
  181. if (!isexist)
  182. {
  183. rdlist = RT_NULL;
  184. rdlist = (struct rt_ready_list *)rt_malloc(sizeof(struct rt_ready_list));
  185. if (rdlist != RT_NULL)
  186. {
  187. rdlist->rdl_event = fdl;
  188. rdlist->rdl_event->epev.events = fdl->revents & revents;
  189. rdlist->next = ep->rdlist->next;
  190. rdlist->exclusive = 0;
  191. ep->rdlist->next = rdlist;
  192. ep->eventpoll_num ++;
  193. res = 0;
  194. if (rdlist->rdl_event->revents & EPOLLONESHOT)
  195. {
  196. rdlist->rdl_event->revents = 0;
  197. }
  198. }
  199. }
  200. ep->tirggered = 1;
  201. rt_mutex_release(&ep->lock);
  202. return res;
  203. }
  204. static int epoll_wqueue_callback(struct rt_wqueue_node *wait, void *key)
  205. {
  206. struct rt_fd_list *fdlist;
  207. if (key && !((rt_ubase_t)key & wait->key))
  208. return -1;
  209. fdlist = rt_container_of(wait, struct rt_fd_list, wqn);
  210. if (fdlist->revents)
  211. {
  212. epoll_rdlist_add(fdlist, (rt_ubase_t)key);
  213. }
  214. return __wqueue_default_wake(wait, key);
  215. }
  216. static void epoll_wqueue_add_callback(rt_wqueue_t *wq, rt_pollreq_t *req)
  217. {
  218. struct rt_fd_list *fdlist;
  219. struct rt_eventpoll *ep;
  220. fdlist = rt_container_of(req, struct rt_fd_list, req);
  221. ep = fdlist->ep;
  222. fdlist->wqn.key = req->_key;
  223. rt_list_init(&(fdlist->wqn.list));
  224. fdlist->wqn.polling_thread = ep->polling_thread;
  225. fdlist->wqn.wakeup = epoll_wqueue_callback;
  226. rt_wqueue_add(wq, &fdlist->wqn);
  227. }
  228. static void epoll_ctl_install(struct rt_fd_list *fdlist, struct rt_eventpoll *ep)
  229. {
  230. rt_uint32_t mask = 0;
  231. fdlist->req._key = fdlist->revents;
  232. mask = epoll_get_event(fdlist, &fdlist->req);
  233. if (mask & fdlist->revents)
  234. {
  235. epoll_rdlist_add(fdlist, mask);
  236. }
  237. }
  238. static void epoll_member_init(struct rt_eventpoll *ep)
  239. {
  240. ep->tirggered = 0;
  241. ep->eventpoll_num = 0;
  242. ep->polling_thread = rt_thread_self();
  243. ep->rdlist = RT_NULL;
  244. ep->fdlist = RT_NULL;
  245. ep->req._key = 0;
  246. rt_wqueue_init(&ep->epoll_read);
  247. rt_spin_lock_init(&spinlock);
  248. }
  249. static int epoll_epf_init(int fd)
  250. {
  251. struct dfs_file *df;
  252. struct rt_eventpoll *ep;
  253. rt_err_t ret = 0;
  254. df = fd_get(fd);
  255. if (df)
  256. {
  257. ep = (struct rt_eventpoll *)rt_malloc(sizeof(struct rt_eventpoll));
  258. if (ep)
  259. {
  260. epoll_member_init(ep);
  261. rt_mutex_init(&ep->lock, EPOLL_MUTEX_NAME, RT_IPC_FLAG_FIFO);
  262. #ifdef RT_USING_DFS_V2
  263. df->fops = &epoll_fops;
  264. #endif
  265. df->vnode = (struct dfs_vnode *)rt_malloc(sizeof(struct dfs_vnode));
  266. if (df->vnode)
  267. {
  268. ep->fdlist = (struct rt_fd_list *)rt_malloc(sizeof(struct rt_fd_list));
  269. if (ep->fdlist)
  270. {
  271. ep->fdlist->next = RT_NULL;
  272. ep->fdlist->epev.data.fd = fd;
  273. ep->fdlist->ep = ep;
  274. dfs_vnode_init(df->vnode, FT_REGULAR, &epoll_fops);
  275. df->vnode->data = ep;
  276. }
  277. else
  278. {
  279. ret = -ENOMEM;
  280. rt_free(df->vnode);
  281. rt_free(ep);
  282. }
  283. }
  284. else
  285. {
  286. ret = -ENOMEM;
  287. rt_free(ep);
  288. }
  289. }
  290. else
  291. {
  292. ret = -ENOMEM;
  293. }
  294. }
  295. return ret;
  296. }
  297. static int epoll_do_create(int size)
  298. {
  299. rt_err_t ret = -1;
  300. int status;
  301. int fd;
  302. if (size < 0)
  303. {
  304. rt_set_errno(EINVAL);
  305. }
  306. else
  307. {
  308. fd = fd_new();
  309. if (fd >= 0)
  310. {
  311. ret = fd;
  312. status = epoll_epf_init(fd);
  313. if (status < 0)
  314. {
  315. fd_release(fd);
  316. rt_set_errno(-status);
  317. }
  318. }
  319. else
  320. {
  321. rt_set_errno(-fd);
  322. }
  323. }
  324. return ret;
  325. }
  326. static int epoll_ctl_add(struct dfs_file *df ,struct epoll_event *event)
  327. {
  328. struct rt_fd_list *fdlist;
  329. struct rt_eventpoll *ep;
  330. rt_err_t ret = -EINVAL;
  331. if (df->vnode->data)
  332. {
  333. ep = df->vnode->data;
  334. fdlist = ep->fdlist;
  335. ret = 0;
  336. while (fdlist->next != RT_NULL)
  337. {
  338. if (fdlist->next->epev.data.fd == event->data.fd)
  339. {
  340. return 0;
  341. }
  342. fdlist = fdlist->next;
  343. }
  344. fdlist = (struct rt_fd_list *)rt_malloc(sizeof(struct rt_fd_list));
  345. if (fdlist)
  346. {
  347. fdlist->epev.data.fd = event->data.fd;
  348. fdlist->epev.events = event->events;
  349. fdlist->ep = ep;
  350. fdlist->req._proc = epoll_wqueue_add_callback;
  351. fdlist->next = ep->fdlist->next;
  352. fdlist->revents = event->events;
  353. ep->fdlist->next = fdlist;
  354. epoll_ctl_install(fdlist, ep);
  355. }
  356. else
  357. {
  358. ret = -ENOMEM;
  359. }
  360. }
  361. return ret;
  362. }
  363. static int epoll_ctl_del(struct dfs_file *df ,struct epoll_event *event)
  364. {
  365. struct rt_fd_list *fdlist, *fre_fd;
  366. struct rt_eventpoll *ep = RT_NULL;
  367. struct rt_ready_list *rdlist, *fre_rdl;
  368. rt_err_t ret = -EINVAL;
  369. if (df->vnode->data)
  370. {
  371. ep = df->vnode->data;
  372. fdlist = ep->fdlist;
  373. while (fdlist->next != RT_NULL)
  374. {
  375. if (fdlist->next->epev.data.fd == event->data.fd)
  376. {
  377. fre_fd = fdlist->next;
  378. fdlist->next = fdlist->next->next;
  379. if (fre_fd->epev.events != 0)
  380. {
  381. rt_wqueue_remove(&fre_fd->wqn);
  382. }
  383. rt_free(fre_fd);
  384. break;
  385. }
  386. else
  387. {
  388. fdlist = fdlist->next;
  389. }
  390. }
  391. if (ep->rdlist)
  392. {
  393. rdlist = ep->rdlist;
  394. while (rdlist->next != RT_NULL)
  395. {
  396. if (rdlist->next->rdl_event->epev.data.fd == event->data.fd)
  397. {
  398. fre_rdl = rdlist->next;
  399. rdlist->next = rdlist->next->next;
  400. ep->eventpoll_num --;
  401. rt_free(fre_rdl);
  402. break;
  403. }
  404. else
  405. {
  406. rdlist = rdlist->next;
  407. }
  408. }
  409. }
  410. ret = 0;
  411. }
  412. return ret;
  413. }
  414. static int epoll_ctl_mod(struct dfs_file *df ,struct epoll_event *event)
  415. {
  416. struct rt_fd_list *fdlist;
  417. struct rt_eventpoll *ep = RT_NULL;
  418. rt_err_t ret = -EINVAL;
  419. if (df->vnode->data)
  420. {
  421. ep = df->vnode->data;
  422. fdlist = ep->fdlist;
  423. while (fdlist->next != RT_NULL)
  424. {
  425. if (fdlist->next->epev.data.fd == event->data.fd)
  426. {
  427. fdlist->next->epev.events = event->events;
  428. fdlist->next->revents = event->events;
  429. rt_wqueue_remove(&fdlist->next->wqn);
  430. epoll_ctl_install(fdlist->next, ep);
  431. break;
  432. }
  433. fdlist = fdlist->next;
  434. }
  435. ret = 0;
  436. }
  437. return ret;
  438. }
  439. static int epoll_do_ctl(int epfd, int op, int fd, struct epoll_event *event)
  440. {
  441. struct dfs_file *epdf;
  442. struct rt_eventpoll *ep;
  443. rt_err_t ret = 0;
  444. if (op & ~EFD_SHARED_EPOLL_TYPE)
  445. {
  446. rt_set_errno(EINVAL);
  447. return -1;
  448. }
  449. if ((epfd == fd) || (epfd < 0) || (fd < 0) || (event->data.fd != fd))
  450. {
  451. rt_set_errno(EINVAL);
  452. return -1;
  453. }
  454. if (!(event->events & EPOLLEXCLUSIVE_BITS))
  455. {
  456. rt_set_errno(EINVAL);
  457. return -1;
  458. }
  459. epdf = fd_get(epfd);
  460. if (epdf->vnode->data)
  461. {
  462. ep = epdf->vnode->data;
  463. rt_mutex_take(&ep->lock, RT_WAITING_FOREVER);
  464. switch (op)
  465. {
  466. case EPOLL_CTL_ADD:
  467. ret = epoll_ctl_add(epdf, event);
  468. break;
  469. case EPOLL_CTL_DEL:
  470. ret = epoll_ctl_del(epdf, event);
  471. break;
  472. case EPOLL_CTL_MOD:
  473. ret = epoll_ctl_mod(epdf, event);
  474. break;
  475. default:
  476. rt_set_errno(EINVAL);
  477. break;
  478. }
  479. if (ret < 0)
  480. {
  481. rt_set_errno(-ret);
  482. ret = -1;
  483. }
  484. rt_mutex_release(&ep->lock);
  485. }
  486. return ret;
  487. }
  488. static int epoll_wait_timeout(struct rt_eventpoll *ep, int msec)
  489. {
  490. rt_int32_t timeout;
  491. struct rt_thread *thread;
  492. rt_base_t level;
  493. int ret = 0;
  494. thread = ep->polling_thread;
  495. timeout = rt_tick_from_millisecond(msec);
  496. level = rt_spin_lock_irqsave(&spinlock);
  497. if (timeout != 0 && !ep->tirggered)
  498. {
  499. if (rt_thread_suspend_with_flag(thread, RT_KILLABLE) == RT_EOK)
  500. {
  501. if (timeout > 0)
  502. {
  503. rt_timer_control(&(thread->thread_timer),
  504. RT_TIMER_CTRL_SET_TIME,
  505. &timeout);
  506. rt_timer_start(&(thread->thread_timer));
  507. }
  508. rt_spin_unlock_irqrestore(&spinlock, level);
  509. rt_schedule();
  510. level = rt_spin_lock_irqsave(&spinlock);
  511. }
  512. }
  513. ret = !ep->tirggered;
  514. rt_spin_unlock_irqrestore(&spinlock, level);
  515. return ret;
  516. }
  517. static int epoll_get_event(struct rt_fd_list *fl, rt_pollreq_t *req)
  518. {
  519. struct dfs_file *df;
  520. int mask = 0;
  521. int fd = 0;
  522. fd = fl->epev.data.fd;
  523. if (fd >= 0)
  524. {
  525. df = fd_get(fd);
  526. if (df)
  527. {
  528. if (df->vnode->fops->poll)
  529. {
  530. req->_key = fl->revents | POLLERR | POLLHUP;
  531. mask = df->vnode->fops->poll(df, req);
  532. if (mask < 0)
  533. return mask;
  534. }
  535. mask &= fl->revents | EPOLLOUT | POLLERR;
  536. }
  537. }
  538. return mask;
  539. }
  540. static int epoll_do(struct rt_eventpoll *ep, struct epoll_event *events, int maxevents, int timeout)
  541. {
  542. struct rt_ready_list *rdlist, *pre_rdlist;
  543. int event_num = 0;
  544. int istimeout = 0;
  545. int isn_add = 0;
  546. int isfree = 0;
  547. int mask = 0;
  548. while (1)
  549. {
  550. rt_mutex_take(&ep->lock, RT_WAITING_FOREVER);
  551. if (ep->eventpoll_num > 0)
  552. {
  553. rdlist = ep->rdlist;
  554. while (rdlist->next != RT_NULL)
  555. {
  556. isfree = 0;
  557. isn_add = 0;
  558. pre_rdlist = rdlist;
  559. rdlist = rdlist->next;
  560. if (event_num < maxevents)
  561. {
  562. if (rdlist->rdl_event->revents == 0)
  563. {
  564. isfree = 1;
  565. rt_wqueue_remove(&rdlist->rdl_event->wqn);
  566. }
  567. else
  568. {
  569. if (rdlist->rdl_event->revents & EPOLLET)
  570. {
  571. rt_wqueue_remove(&rdlist->rdl_event->wqn);
  572. mask = epoll_get_event(rdlist->rdl_event, &rdlist->rdl_event->req);
  573. rdlist->rdl_event->epev.events = mask & rdlist->rdl_event->revents;
  574. isfree = 1;
  575. }
  576. else
  577. {
  578. if (rdlist->exclusive)
  579. {
  580. rt_wqueue_remove(&rdlist->rdl_event->wqn);
  581. mask = epoll_get_event(rdlist->rdl_event, &rdlist->rdl_event->req);
  582. if (mask & rdlist->rdl_event->revents)
  583. {
  584. rdlist->rdl_event->epev.events = mask;
  585. }
  586. else
  587. {
  588. isfree = 1;
  589. isn_add = 1;
  590. }
  591. }
  592. else
  593. {
  594. rdlist->exclusive = 1;
  595. }
  596. }
  597. }
  598. if (!isn_add)
  599. {
  600. events[event_num].data.fd = rdlist->rdl_event->epev.data.fd;
  601. events[event_num].events = rdlist->rdl_event->epev.events;
  602. event_num ++;
  603. }
  604. if (isfree)
  605. {
  606. pre_rdlist->next = rdlist->next;
  607. rt_free(rdlist);
  608. ep->eventpoll_num --;
  609. rdlist = pre_rdlist;
  610. }
  611. }
  612. else
  613. {
  614. break;
  615. }
  616. }
  617. }
  618. rt_mutex_release(&ep->lock);
  619. if (event_num || istimeout)
  620. {
  621. ep->tirggered = 0;
  622. break;
  623. }
  624. if (epoll_wait_timeout(ep, timeout))
  625. {
  626. istimeout = 1;
  627. }
  628. }
  629. return event_num;
  630. }
  631. static int epoll_do_wait(int epfd, struct epoll_event *events, int maxevents, int timeout, const sigset_t *ss)
  632. {
  633. struct rt_eventpoll *ep;
  634. struct dfs_file *df;
  635. lwp_sigset_t old_sig, new_sig;
  636. rt_err_t ret = 0;
  637. if (ss)
  638. {
  639. memcpy(&new_sig, ss, sizeof(lwp_sigset_t));
  640. lwp_thread_signal_mask(rt_thread_self(), LWP_SIG_MASK_CMD_BLOCK, &new_sig, &old_sig);
  641. }
  642. if ((maxevents > 0) && (epfd >=0))
  643. {
  644. df = fd_get(epfd);
  645. if (df && df->vnode)
  646. {
  647. ep = (struct rt_eventpoll *)df->vnode->data;
  648. if (ep)
  649. {
  650. ret = epoll_do(ep, events, maxevents, timeout);
  651. }
  652. }
  653. }
  654. if (ss)
  655. {
  656. lwp_thread_signal_mask(rt_thread_self(), LWP_SIG_MASK_CMD_SET_MASK, &old_sig, RT_NULL);
  657. }
  658. if (ret < 0)
  659. {
  660. rt_set_errno(-ret);
  661. ret = -1;
  662. }
  663. return ret;
  664. }
  665. int epoll_create(int size)
  666. {
  667. return epoll_do_create(size);
  668. }
  669. int epoll_ctl(int epfd, int op, int fd, struct epoll_event *event)
  670. {
  671. return epoll_do_ctl(epfd, op, fd, event);
  672. }
  673. int epoll_wait(int epfd, struct epoll_event *events, int maxevents, int timeout)
  674. {
  675. return epoll_do_wait(epfd, events, maxevents, timeout, RT_NULL);
  676. }
  677. int epoll_pwait(int epfd, struct epoll_event *events, int maxevents, int timeout, const sigset_t *ss)
  678. {
  679. return epoll_do_wait(epfd, events, maxevents, timeout, ss);
  680. }
  681. int epoll_pwait2(int epfd, struct epoll_event *events, int maxevents, int timeout, const sigset_t *ss)
  682. {
  683. return epoll_do_wait(epfd, events, maxevents, timeout, ss);
  684. }