mirror of
https://github.com/torvalds/linux.git
synced 2024-11-25 05:32:00 +00:00
df9b455633
In netfslib, a buffered writeback operation has a 'write queue' of folios
that are being written, held in a linear sequence of folio_queue structs.
The 'issuer' adds new folio_queues on the leading edge of the queue and
populates each one progressively; the 'collector' pops them off the
trailing edge and discards them and the folios they point to as they are
consumed.
The queue is required to always retain at least one folio_queue structure.
This allows the queue to be accessed without locking and with just a bit of
barriering.
When a new subrequest is prepared, its ->io_iter iterator is pointed at the
current end of the write queue and then the iterator is extended as more
data is added to the queue until the subrequest is committed.
Now, the problem is that the folio_queue at the leading edge of the write
queue when a subrequest is prepared might have been entirely consumed - but
not yet removed from the queue as it is the only remaining one and is
preventing the queue from collapsing.
So, what happens is that subreq->io_iter is pointed at the spent
folio_queue, then a new folio_queue is added, and, at that point, the
collector is at entirely at liberty to immediately delete the spent
folio_queue.
This leaves the subreq->io_iter pointing at a freed object. If the system
is lucky, iterate_folioq() sees ->io_iter, sees the as-yet uncorrupted
freed object and advances to the next folio_queue in the queue.
In the case seen, however, the freed object gets recycled and put back onto
the queue at the tail and filled to the end. This confuses
iterate_folioq() and it tries to step ->next, which may be NULL - resulting
in an oops.
Fix this by the following means:
(1) When preparing a write subrequest, make sure there's a folio_queue
struct with space in it at the leading edge of the queue. A function
to make space is split out of the function to append a folio so that
it can be called for this purpose.
(2) If the request struct iterator is pointing to a completely spent
folio_queue when we make space, then advance the iterator to the newly
allocated folio_queue. The subrequest's iterator will then be set
from this.
The oops could be triggered using the generic/346 xfstest with a filesystem
on9P over TCP with cache=loose. The oops looked something like:
BUG: kernel NULL pointer dereference, address: 0000000000000008
#PF: supervisor read access in kernel mode
#PF: error_code(0x0000) - not-present page
...
RIP: 0010:_copy_from_iter+0x2db/0x530
...
Call Trace:
<TASK>
...
p9pdu_vwritef+0x3d8/0x5d0
p9_client_prepare_req+0xa8/0x140
p9_client_rpc+0x81/0x280
p9_client_write+0xcf/0x1c0
v9fs_issue_write+0x87/0xc0
netfs_advance_write+0xa0/0xb0
netfs_write_folio.isra.0+0x42d/0x500
netfs_writepages+0x15a/0x1f0
do_writepages+0xd1/0x220
filemap_fdatawrite_wbc+0x5c/0x80
v9fs_mmap_vm_close+0x7d/0xb0
remove_vma+0x35/0x70
vms_complete_munmap_vmas+0x11a/0x170
do_vmi_align_munmap+0x17d/0x1c0
do_vmi_munmap+0x13e/0x150
__vm_munmap+0x92/0xd0
__x64_sys_munmap+0x17/0x20
do_syscall_64+0x80/0xe0
entry_SYSCALL_64_after_hwframe+0x71/0x79
This also fixed a similar-looking issue with cifs and generic/074.
Fixes: cd0277ed0c
("netfs: Use new folio_queue data type and iterator instead of xarray iter")
Reported-by: kernel test robot <oliver.sang@intel.com>
Closes: https://lore.kernel.org/oe-lkp/202409180928.f20b5a08-oliver.sang@intel.com
Closes: https://lore.kernel.org/oe-lkp/202409131438.3f225fbf-oliver.sang@intel.com
Signed-off-by: David Howells <dhowells@redhat.com>
Tested-by: kernel test robot <oliver.sang@intel.com>
cc: Eric Van Hensbergen <ericvh@kernel.org>
cc: Latchesar Ionkov <lucho@ionkov.net>
cc: Dominique Martinet <asmadeus@codewreck.org>
cc: Christian Schoenebeck <linux_oss@crudebyte.com>
cc: Paulo Alcantara <pc@manguebit.com>
cc: Jeff Layton <jlayton@kernel.org>
cc: v9fs@lists.linux.dev
cc: linux-cifs@vger.kernel.org
cc: netfs@lists.linux.dev
cc: linux-fsdevel@vger.kernel.org
Signed-off-by: Steve French <stfrench@microsoft.com>
476 lines
13 KiB
C
476 lines
13 KiB
C
/* SPDX-License-Identifier: GPL-2.0-or-later */
|
|
/* Internal definitions for network filesystem support
|
|
*
|
|
* Copyright (C) 2021 Red Hat, Inc. All Rights Reserved.
|
|
* Written by David Howells (dhowells@redhat.com)
|
|
*/
|
|
|
|
#include <linux/slab.h>
|
|
#include <linux/seq_file.h>
|
|
#include <linux/folio_queue.h>
|
|
#include <linux/netfs.h>
|
|
#include <linux/fscache.h>
|
|
#include <linux/fscache-cache.h>
|
|
#include <trace/events/netfs.h>
|
|
#include <trace/events/fscache.h>
|
|
|
|
#ifdef pr_fmt
|
|
#undef pr_fmt
|
|
#endif
|
|
|
|
#define pr_fmt(fmt) "netfs: " fmt
|
|
|
|
/*
|
|
* buffered_read.c
|
|
*/
|
|
int netfs_prefetch_for_write(struct file *file, struct folio *folio,
|
|
size_t offset, size_t len);
|
|
|
|
/*
|
|
* main.c
|
|
*/
|
|
extern unsigned int netfs_debug;
|
|
extern struct list_head netfs_io_requests;
|
|
extern spinlock_t netfs_proc_lock;
|
|
extern mempool_t netfs_request_pool;
|
|
extern mempool_t netfs_subrequest_pool;
|
|
|
|
#ifdef CONFIG_PROC_FS
|
|
static inline void netfs_proc_add_rreq(struct netfs_io_request *rreq)
|
|
{
|
|
spin_lock(&netfs_proc_lock);
|
|
list_add_tail_rcu(&rreq->proc_link, &netfs_io_requests);
|
|
spin_unlock(&netfs_proc_lock);
|
|
}
|
|
static inline void netfs_proc_del_rreq(struct netfs_io_request *rreq)
|
|
{
|
|
if (!list_empty(&rreq->proc_link)) {
|
|
spin_lock(&netfs_proc_lock);
|
|
list_del_rcu(&rreq->proc_link);
|
|
spin_unlock(&netfs_proc_lock);
|
|
}
|
|
}
|
|
#else
|
|
static inline void netfs_proc_add_rreq(struct netfs_io_request *rreq) {}
|
|
static inline void netfs_proc_del_rreq(struct netfs_io_request *rreq) {}
|
|
#endif
|
|
|
|
/*
|
|
* misc.c
|
|
*/
|
|
struct folio_queue *netfs_buffer_make_space(struct netfs_io_request *rreq);
|
|
int netfs_buffer_append_folio(struct netfs_io_request *rreq, struct folio *folio,
|
|
bool needs_put);
|
|
struct folio_queue *netfs_delete_buffer_head(struct netfs_io_request *wreq);
|
|
void netfs_clear_buffer(struct netfs_io_request *rreq);
|
|
void netfs_reset_iter(struct netfs_io_subrequest *subreq);
|
|
|
|
/*
|
|
* objects.c
|
|
*/
|
|
struct netfs_io_request *netfs_alloc_request(struct address_space *mapping,
|
|
struct file *file,
|
|
loff_t start, size_t len,
|
|
enum netfs_io_origin origin);
|
|
void netfs_get_request(struct netfs_io_request *rreq, enum netfs_rreq_ref_trace what);
|
|
void netfs_clear_subrequests(struct netfs_io_request *rreq, bool was_async);
|
|
void netfs_put_request(struct netfs_io_request *rreq, bool was_async,
|
|
enum netfs_rreq_ref_trace what);
|
|
struct netfs_io_subrequest *netfs_alloc_subrequest(struct netfs_io_request *rreq);
|
|
|
|
static inline void netfs_see_request(struct netfs_io_request *rreq,
|
|
enum netfs_rreq_ref_trace what)
|
|
{
|
|
trace_netfs_rreq_ref(rreq->debug_id, refcount_read(&rreq->ref), what);
|
|
}
|
|
|
|
/*
|
|
* read_collect.c
|
|
*/
|
|
void netfs_read_termination_worker(struct work_struct *work);
|
|
void netfs_rreq_terminated(struct netfs_io_request *rreq, bool was_async);
|
|
|
|
/*
|
|
* read_pgpriv2.c
|
|
*/
|
|
void netfs_pgpriv2_mark_copy_to_cache(struct netfs_io_subrequest *subreq,
|
|
struct netfs_io_request *rreq,
|
|
struct folio_queue *folioq,
|
|
int slot);
|
|
void netfs_pgpriv2_write_to_the_cache(struct netfs_io_request *rreq);
|
|
bool netfs_pgpriv2_unlock_copied_folios(struct netfs_io_request *wreq);
|
|
|
|
/*
|
|
* read_retry.c
|
|
*/
|
|
void netfs_retry_reads(struct netfs_io_request *rreq);
|
|
void netfs_unlock_abandoned_read_pages(struct netfs_io_request *rreq);
|
|
|
|
/*
|
|
* stats.c
|
|
*/
|
|
#ifdef CONFIG_NETFS_STATS
|
|
extern atomic_t netfs_n_rh_dio_read;
|
|
extern atomic_t netfs_n_rh_readahead;
|
|
extern atomic_t netfs_n_rh_read_folio;
|
|
extern atomic_t netfs_n_rh_rreq;
|
|
extern atomic_t netfs_n_rh_sreq;
|
|
extern atomic_t netfs_n_rh_download;
|
|
extern atomic_t netfs_n_rh_download_done;
|
|
extern atomic_t netfs_n_rh_download_failed;
|
|
extern atomic_t netfs_n_rh_download_instead;
|
|
extern atomic_t netfs_n_rh_read;
|
|
extern atomic_t netfs_n_rh_read_done;
|
|
extern atomic_t netfs_n_rh_read_failed;
|
|
extern atomic_t netfs_n_rh_zero;
|
|
extern atomic_t netfs_n_rh_short_read;
|
|
extern atomic_t netfs_n_rh_write;
|
|
extern atomic_t netfs_n_rh_write_begin;
|
|
extern atomic_t netfs_n_rh_write_done;
|
|
extern atomic_t netfs_n_rh_write_failed;
|
|
extern atomic_t netfs_n_rh_write_zskip;
|
|
extern atomic_t netfs_n_wh_buffered_write;
|
|
extern atomic_t netfs_n_wh_writethrough;
|
|
extern atomic_t netfs_n_wh_dio_write;
|
|
extern atomic_t netfs_n_wh_writepages;
|
|
extern atomic_t netfs_n_wh_copy_to_cache;
|
|
extern atomic_t netfs_n_wh_wstream_conflict;
|
|
extern atomic_t netfs_n_wh_upload;
|
|
extern atomic_t netfs_n_wh_upload_done;
|
|
extern atomic_t netfs_n_wh_upload_failed;
|
|
extern atomic_t netfs_n_wh_write;
|
|
extern atomic_t netfs_n_wh_write_done;
|
|
extern atomic_t netfs_n_wh_write_failed;
|
|
extern atomic_t netfs_n_wb_lock_skip;
|
|
extern atomic_t netfs_n_wb_lock_wait;
|
|
extern atomic_t netfs_n_folioq;
|
|
|
|
int netfs_stats_show(struct seq_file *m, void *v);
|
|
|
|
static inline void netfs_stat(atomic_t *stat)
|
|
{
|
|
atomic_inc(stat);
|
|
}
|
|
|
|
static inline void netfs_stat_d(atomic_t *stat)
|
|
{
|
|
atomic_dec(stat);
|
|
}
|
|
|
|
#else
|
|
#define netfs_stat(x) do {} while(0)
|
|
#define netfs_stat_d(x) do {} while(0)
|
|
#endif
|
|
|
|
/*
|
|
* write_collect.c
|
|
*/
|
|
int netfs_folio_written_back(struct folio *folio);
|
|
void netfs_write_collection_worker(struct work_struct *work);
|
|
void netfs_wake_write_collector(struct netfs_io_request *wreq, bool was_async);
|
|
|
|
/*
|
|
* write_issue.c
|
|
*/
|
|
struct netfs_io_request *netfs_create_write_req(struct address_space *mapping,
|
|
struct file *file,
|
|
loff_t start,
|
|
enum netfs_io_origin origin);
|
|
void netfs_reissue_write(struct netfs_io_stream *stream,
|
|
struct netfs_io_subrequest *subreq,
|
|
struct iov_iter *source);
|
|
void netfs_issue_write(struct netfs_io_request *wreq,
|
|
struct netfs_io_stream *stream);
|
|
int netfs_advance_write(struct netfs_io_request *wreq,
|
|
struct netfs_io_stream *stream,
|
|
loff_t start, size_t len, bool to_eof);
|
|
struct netfs_io_request *netfs_begin_writethrough(struct kiocb *iocb, size_t len);
|
|
int netfs_advance_writethrough(struct netfs_io_request *wreq, struct writeback_control *wbc,
|
|
struct folio *folio, size_t copied, bool to_page_end,
|
|
struct folio **writethrough_cache);
|
|
int netfs_end_writethrough(struct netfs_io_request *wreq, struct writeback_control *wbc,
|
|
struct folio *writethrough_cache);
|
|
int netfs_unbuffered_write(struct netfs_io_request *wreq, bool may_wait, size_t len);
|
|
|
|
/*
|
|
* Miscellaneous functions.
|
|
*/
|
|
static inline bool netfs_is_cache_enabled(struct netfs_inode *ctx)
|
|
{
|
|
#if IS_ENABLED(CONFIG_FSCACHE)
|
|
struct fscache_cookie *cookie = ctx->cache;
|
|
|
|
return fscache_cookie_valid(cookie) && cookie->cache_priv &&
|
|
fscache_cookie_enabled(cookie);
|
|
#else
|
|
return false;
|
|
#endif
|
|
}
|
|
|
|
/*
|
|
* Get a ref on a netfs group attached to a dirty page (e.g. a ceph snap).
|
|
*/
|
|
static inline struct netfs_group *netfs_get_group(struct netfs_group *netfs_group)
|
|
{
|
|
if (netfs_group && netfs_group != NETFS_FOLIO_COPY_TO_CACHE)
|
|
refcount_inc(&netfs_group->ref);
|
|
return netfs_group;
|
|
}
|
|
|
|
/*
|
|
* Dispose of a netfs group attached to a dirty page (e.g. a ceph snap).
|
|
*/
|
|
static inline void netfs_put_group(struct netfs_group *netfs_group)
|
|
{
|
|
if (netfs_group &&
|
|
netfs_group != NETFS_FOLIO_COPY_TO_CACHE &&
|
|
refcount_dec_and_test(&netfs_group->ref))
|
|
netfs_group->free(netfs_group);
|
|
}
|
|
|
|
/*
|
|
* Dispose of a netfs group attached to a dirty page (e.g. a ceph snap).
|
|
*/
|
|
static inline void netfs_put_group_many(struct netfs_group *netfs_group, int nr)
|
|
{
|
|
if (netfs_group &&
|
|
netfs_group != NETFS_FOLIO_COPY_TO_CACHE &&
|
|
refcount_sub_and_test(nr, &netfs_group->ref))
|
|
netfs_group->free(netfs_group);
|
|
}
|
|
|
|
/*
|
|
* fscache-cache.c
|
|
*/
|
|
#ifdef CONFIG_PROC_FS
|
|
extern const struct seq_operations fscache_caches_seq_ops;
|
|
#endif
|
|
bool fscache_begin_cache_access(struct fscache_cache *cache, enum fscache_access_trace why);
|
|
void fscache_end_cache_access(struct fscache_cache *cache, enum fscache_access_trace why);
|
|
struct fscache_cache *fscache_lookup_cache(const char *name, bool is_cache);
|
|
void fscache_put_cache(struct fscache_cache *cache, enum fscache_cache_trace where);
|
|
|
|
static inline enum fscache_cache_state fscache_cache_state(const struct fscache_cache *cache)
|
|
{
|
|
return smp_load_acquire(&cache->state);
|
|
}
|
|
|
|
static inline bool fscache_cache_is_live(const struct fscache_cache *cache)
|
|
{
|
|
return fscache_cache_state(cache) == FSCACHE_CACHE_IS_ACTIVE;
|
|
}
|
|
|
|
static inline void fscache_set_cache_state(struct fscache_cache *cache,
|
|
enum fscache_cache_state new_state)
|
|
{
|
|
smp_store_release(&cache->state, new_state);
|
|
|
|
}
|
|
|
|
static inline bool fscache_set_cache_state_maybe(struct fscache_cache *cache,
|
|
enum fscache_cache_state old_state,
|
|
enum fscache_cache_state new_state)
|
|
{
|
|
return try_cmpxchg_release(&cache->state, &old_state, new_state);
|
|
}
|
|
|
|
/*
|
|
* fscache-cookie.c
|
|
*/
|
|
extern struct kmem_cache *fscache_cookie_jar;
|
|
#ifdef CONFIG_PROC_FS
|
|
extern const struct seq_operations fscache_cookies_seq_ops;
|
|
#endif
|
|
extern struct timer_list fscache_cookie_lru_timer;
|
|
|
|
extern void fscache_print_cookie(struct fscache_cookie *cookie, char prefix);
|
|
extern bool fscache_begin_cookie_access(struct fscache_cookie *cookie,
|
|
enum fscache_access_trace why);
|
|
|
|
static inline void fscache_see_cookie(struct fscache_cookie *cookie,
|
|
enum fscache_cookie_trace where)
|
|
{
|
|
trace_fscache_cookie(cookie->debug_id, refcount_read(&cookie->ref),
|
|
where);
|
|
}
|
|
|
|
/*
|
|
* fscache-main.c
|
|
*/
|
|
extern unsigned int fscache_hash(unsigned int salt, const void *data, size_t len);
|
|
#ifdef CONFIG_FSCACHE
|
|
int __init fscache_init(void);
|
|
void __exit fscache_exit(void);
|
|
#else
|
|
static inline int fscache_init(void) { return 0; }
|
|
static inline void fscache_exit(void) {}
|
|
#endif
|
|
|
|
/*
|
|
* fscache-proc.c
|
|
*/
|
|
#ifdef CONFIG_PROC_FS
|
|
extern int __init fscache_proc_init(void);
|
|
extern void fscache_proc_cleanup(void);
|
|
#else
|
|
#define fscache_proc_init() (0)
|
|
#define fscache_proc_cleanup() do {} while (0)
|
|
#endif
|
|
|
|
/*
|
|
* fscache-stats.c
|
|
*/
|
|
#ifdef CONFIG_FSCACHE_STATS
|
|
extern atomic_t fscache_n_volumes;
|
|
extern atomic_t fscache_n_volumes_collision;
|
|
extern atomic_t fscache_n_volumes_nomem;
|
|
extern atomic_t fscache_n_cookies;
|
|
extern atomic_t fscache_n_cookies_lru;
|
|
extern atomic_t fscache_n_cookies_lru_expired;
|
|
extern atomic_t fscache_n_cookies_lru_removed;
|
|
extern atomic_t fscache_n_cookies_lru_dropped;
|
|
|
|
extern atomic_t fscache_n_acquires;
|
|
extern atomic_t fscache_n_acquires_ok;
|
|
extern atomic_t fscache_n_acquires_oom;
|
|
|
|
extern atomic_t fscache_n_invalidates;
|
|
|
|
extern atomic_t fscache_n_relinquishes;
|
|
extern atomic_t fscache_n_relinquishes_retire;
|
|
extern atomic_t fscache_n_relinquishes_dropped;
|
|
|
|
extern atomic_t fscache_n_resizes;
|
|
extern atomic_t fscache_n_resizes_null;
|
|
|
|
static inline void fscache_stat(atomic_t *stat)
|
|
{
|
|
atomic_inc(stat);
|
|
}
|
|
|
|
static inline void fscache_stat_d(atomic_t *stat)
|
|
{
|
|
atomic_dec(stat);
|
|
}
|
|
|
|
#define __fscache_stat(stat) (stat)
|
|
|
|
int fscache_stats_show(struct seq_file *m);
|
|
#else
|
|
|
|
#define __fscache_stat(stat) (NULL)
|
|
#define fscache_stat(stat) do {} while (0)
|
|
#define fscache_stat_d(stat) do {} while (0)
|
|
|
|
static inline int fscache_stats_show(struct seq_file *m) { return 0; }
|
|
#endif
|
|
|
|
/*
|
|
* fscache-volume.c
|
|
*/
|
|
#ifdef CONFIG_PROC_FS
|
|
extern const struct seq_operations fscache_volumes_seq_ops;
|
|
#endif
|
|
|
|
struct fscache_volume *fscache_get_volume(struct fscache_volume *volume,
|
|
enum fscache_volume_trace where);
|
|
bool fscache_begin_volume_access(struct fscache_volume *volume,
|
|
struct fscache_cookie *cookie,
|
|
enum fscache_access_trace why);
|
|
void fscache_create_volume(struct fscache_volume *volume, bool wait);
|
|
|
|
/*****************************************************************************/
|
|
/*
|
|
* debug tracing
|
|
*/
|
|
#define dbgprintk(FMT, ...) \
|
|
printk("[%-6.6s] "FMT"\n", current->comm, ##__VA_ARGS__)
|
|
|
|
#define kenter(FMT, ...) dbgprintk("==> %s("FMT")", __func__, ##__VA_ARGS__)
|
|
#define kleave(FMT, ...) dbgprintk("<== %s()"FMT"", __func__, ##__VA_ARGS__)
|
|
#define kdebug(FMT, ...) dbgprintk(FMT, ##__VA_ARGS__)
|
|
|
|
#ifdef __KDEBUG
|
|
#define _enter(FMT, ...) kenter(FMT, ##__VA_ARGS__)
|
|
#define _leave(FMT, ...) kleave(FMT, ##__VA_ARGS__)
|
|
#define _debug(FMT, ...) kdebug(FMT, ##__VA_ARGS__)
|
|
|
|
#elif defined(CONFIG_NETFS_DEBUG)
|
|
#define _enter(FMT, ...) \
|
|
do { \
|
|
if (netfs_debug) \
|
|
kenter(FMT, ##__VA_ARGS__); \
|
|
} while (0)
|
|
|
|
#define _leave(FMT, ...) \
|
|
do { \
|
|
if (netfs_debug) \
|
|
kleave(FMT, ##__VA_ARGS__); \
|
|
} while (0)
|
|
|
|
#define _debug(FMT, ...) \
|
|
do { \
|
|
if (netfs_debug) \
|
|
kdebug(FMT, ##__VA_ARGS__); \
|
|
} while (0)
|
|
|
|
#else
|
|
#define _enter(FMT, ...) no_printk("==> %s("FMT")", __func__, ##__VA_ARGS__)
|
|
#define _leave(FMT, ...) no_printk("<== %s()"FMT"", __func__, ##__VA_ARGS__)
|
|
#define _debug(FMT, ...) no_printk(FMT, ##__VA_ARGS__)
|
|
#endif
|
|
|
|
/*
|
|
* assertions
|
|
*/
|
|
#if 1 /* defined(__KDEBUGALL) */
|
|
|
|
#define ASSERT(X) \
|
|
do { \
|
|
if (unlikely(!(X))) { \
|
|
pr_err("\n"); \
|
|
pr_err("Assertion failed\n"); \
|
|
BUG(); \
|
|
} \
|
|
} while (0)
|
|
|
|
#define ASSERTCMP(X, OP, Y) \
|
|
do { \
|
|
if (unlikely(!((X) OP (Y)))) { \
|
|
pr_err("\n"); \
|
|
pr_err("Assertion failed\n"); \
|
|
pr_err("%lx " #OP " %lx is false\n", \
|
|
(unsigned long)(X), (unsigned long)(Y)); \
|
|
BUG(); \
|
|
} \
|
|
} while (0)
|
|
|
|
#define ASSERTIF(C, X) \
|
|
do { \
|
|
if (unlikely((C) && !(X))) { \
|
|
pr_err("\n"); \
|
|
pr_err("Assertion failed\n"); \
|
|
BUG(); \
|
|
} \
|
|
} while (0)
|
|
|
|
#define ASSERTIFCMP(C, X, OP, Y) \
|
|
do { \
|
|
if (unlikely((C) && !((X) OP (Y)))) { \
|
|
pr_err("\n"); \
|
|
pr_err("Assertion failed\n"); \
|
|
pr_err("%lx " #OP " %lx is false\n", \
|
|
(unsigned long)(X), (unsigned long)(Y)); \
|
|
BUG(); \
|
|
} \
|
|
} while (0)
|
|
|
|
#else
|
|
|
|
#define ASSERT(X) do {} while (0)
|
|
#define ASSERTCMP(X, OP, Y) do {} while (0)
|
|
#define ASSERTIF(C, X) do {} while (0)
|
|
#define ASSERTIFCMP(C, X, OP, Y) do {} while (0)
|
|
|
|
#endif /* assert or not */
|