2005-11-15 20:53:48 +00:00
|
|
|
/*
|
|
|
|
* SPU file system
|
|
|
|
*
|
|
|
|
* (C) Copyright IBM Deutschland Entwicklung GmbH 2005
|
|
|
|
*
|
|
|
|
* Author: Arnd Bergmann <arndb@de.ibm.com>
|
|
|
|
*
|
|
|
|
* This program is free software; you can redistribute it and/or modify
|
|
|
|
* it under the terms of the GNU General Public License as published by
|
|
|
|
* the Free Software Foundation; either version 2, or (at your option)
|
|
|
|
* any later version.
|
|
|
|
*
|
|
|
|
* This program is distributed in the hope that it will be useful,
|
|
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
|
|
* GNU General Public License for more details.
|
|
|
|
*
|
|
|
|
* You should have received a copy of the GNU General Public License
|
|
|
|
* along with this program; if not, write to the Free Software
|
|
|
|
* Foundation, Inc., 675 Mass Ave, Cambridge, MA 02139, USA.
|
|
|
|
*/
|
|
|
|
#ifndef SPUFS_H
|
|
|
|
#define SPUFS_H
|
|
|
|
|
|
|
|
#include <linux/kref.h>
|
2007-02-13 20:36:50 +00:00
|
|
|
#include <linux/mutex.h>
|
2005-11-15 20:53:48 +00:00
|
|
|
#include <linux/spinlock.h>
|
|
|
|
#include <linux/fs.h>
|
2007-06-29 00:57:56 +00:00
|
|
|
#include <linux/cpumask.h>
|
2005-11-15 20:53:48 +00:00
|
|
|
|
|
|
|
#include <asm/spu.h>
|
2005-11-15 20:53:49 +00:00
|
|
|
#include <asm/spu_csa.h>
|
2006-11-20 17:44:58 +00:00
|
|
|
#include <asm/spu_info.h>
|
2005-11-15 20:53:48 +00:00
|
|
|
|
|
|
|
/* The magic number for our file system */
|
|
|
|
enum {
|
|
|
|
SPUFS_MAGIC = 0x23c9b64e,
|
|
|
|
};
|
|
|
|
|
2005-11-15 20:53:52 +00:00
|
|
|
struct spu_context_ops;
|
2006-10-04 15:26:15 +00:00
|
|
|
struct spu_gang;
|
|
|
|
|
2007-06-29 00:58:03 +00:00
|
|
|
/*
|
|
|
|
* This is the state for spu utilization reporting to userspace.
|
|
|
|
* Because this state is visible to userspace it must never change and needs
|
|
|
|
* to be kept strictly separate from any internal state kept by the kernel.
|
|
|
|
*/
|
|
|
|
enum spuctx_execution_state {
|
|
|
|
SPUCTX_UTIL_USER = 0,
|
|
|
|
SPUCTX_UTIL_SYSTEM,
|
|
|
|
SPUCTX_UTIL_IOWAIT,
|
|
|
|
SPUCTX_UTIL_LOADED,
|
|
|
|
SPUCTX_UTIL_MAX
|
|
|
|
};
|
|
|
|
|
2005-11-15 20:53:48 +00:00
|
|
|
struct spu_context {
|
|
|
|
struct spu *spu; /* pointer to a physical SPU */
|
2005-11-15 20:53:49 +00:00
|
|
|
struct spu_state csa; /* SPU context save area. */
|
2005-11-15 20:53:48 +00:00
|
|
|
spinlock_t mmio_lock; /* protects mmio access */
|
2006-03-22 23:00:12 +00:00
|
|
|
struct address_space *local_store; /* local store mapping. */
|
|
|
|
struct address_space *mfc; /* 'mfc' area mappings. */
|
2007-04-23 19:08:07 +00:00
|
|
|
struct address_space *cntl; /* 'control' area mappings. */
|
|
|
|
struct address_space *signal1; /* 'signal1' area mappings. */
|
|
|
|
struct address_space *signal2; /* 'signal2' area mappings. */
|
|
|
|
struct address_space *mss; /* 'mss' area mappings. */
|
|
|
|
struct address_space *psmap; /* 'psmap' area mappings. */
|
2007-06-04 13:26:51 +00:00
|
|
|
struct mutex mapping_lock;
|
2006-10-04 15:26:21 +00:00
|
|
|
u64 object_id; /* user space pointer for oprofile */
|
2005-11-15 20:53:52 +00:00
|
|
|
|
|
|
|
enum { SPU_STATE_RUNNABLE, SPU_STATE_SAVED } state;
|
2007-02-13 20:36:50 +00:00
|
|
|
struct mutex state_mutex;
|
2007-04-23 19:08:17 +00:00
|
|
|
struct mutex run_mutex;
|
2005-11-15 20:53:52 +00:00
|
|
|
|
|
|
|
struct mm_struct *owner;
|
2005-11-15 20:53:48 +00:00
|
|
|
|
|
|
|
struct kref kref;
|
2005-11-15 20:53:52 +00:00
|
|
|
wait_queue_head_t ibox_wq;
|
|
|
|
wait_queue_head_t wbox_wq;
|
2005-12-06 03:52:25 +00:00
|
|
|
wait_queue_head_t stop_wq;
|
2006-03-22 23:00:11 +00:00
|
|
|
wait_queue_head_t mfc_wq;
|
2005-11-15 20:53:52 +00:00
|
|
|
struct fasync_struct *ibox_fasync;
|
|
|
|
struct fasync_struct *wbox_fasync;
|
2006-03-22 23:00:11 +00:00
|
|
|
struct fasync_struct *mfc_fasync;
|
|
|
|
u32 tagwait;
|
2005-11-15 20:53:52 +00:00
|
|
|
struct spu_context_ops *ops;
|
2005-12-06 03:52:26 +00:00
|
|
|
struct work_struct reap_work;
|
2006-10-04 15:26:14 +00:00
|
|
|
unsigned long flags;
|
|
|
|
unsigned long event_return;
|
2006-10-04 15:26:15 +00:00
|
|
|
|
|
|
|
struct list_head gang_list;
|
|
|
|
struct spu_gang *gang;
|
2007-02-13 20:54:22 +00:00
|
|
|
|
2007-06-29 00:58:01 +00:00
|
|
|
/* owner thread */
|
|
|
|
pid_t tid;
|
|
|
|
|
2007-02-13 20:54:22 +00:00
|
|
|
/* scheduler fields */
|
2007-06-29 00:58:00 +00:00
|
|
|
struct list_head rq;
|
2007-06-29 00:57:51 +00:00
|
|
|
unsigned int time_slice;
|
2007-02-13 20:54:24 +00:00
|
|
|
unsigned long sched_flags;
|
2007-06-29 00:57:56 +00:00
|
|
|
cpumask_t cpus_allowed;
|
2007-02-13 20:54:29 +00:00
|
|
|
int policy;
|
2007-02-13 20:54:22 +00:00
|
|
|
int prio;
|
2007-06-29 00:58:03 +00:00
|
|
|
|
|
|
|
/* statistics */
|
|
|
|
struct {
|
|
|
|
/* updates protected by ctx->state_mutex */
|
|
|
|
enum spuctx_execution_state execution_state;
|
|
|
|
unsigned long tstamp; /* time of last ctx switch */
|
|
|
|
unsigned long times[SPUCTX_UTIL_MAX];
|
|
|
|
unsigned long long vol_ctx_switch;
|
|
|
|
unsigned long long invol_ctx_switch;
|
|
|
|
unsigned long long min_flt;
|
|
|
|
unsigned long long maj_flt;
|
|
|
|
unsigned long long hash_flt;
|
|
|
|
unsigned long long slb_flt;
|
|
|
|
unsigned long long slb_flt_base; /* # at last ctx switch */
|
|
|
|
unsigned long long class2_intr;
|
|
|
|
unsigned long long class2_intr_base; /* # at last ctx switch */
|
|
|
|
unsigned long long libassist;
|
|
|
|
} stats;
|
2006-10-04 15:26:15 +00:00
|
|
|
};
|
|
|
|
|
|
|
|
struct spu_gang {
|
|
|
|
struct list_head list;
|
|
|
|
struct mutex mutex;
|
|
|
|
struct kref kref;
|
|
|
|
int contexts;
|
2005-11-15 20:53:52 +00:00
|
|
|
};
|
|
|
|
|
2006-03-22 23:00:11 +00:00
|
|
|
struct mfc_dma_command {
|
|
|
|
int32_t pad; /* reserved */
|
|
|
|
uint32_t lsa; /* local storage address */
|
|
|
|
uint64_t ea; /* effective address */
|
|
|
|
uint16_t size; /* transfer size */
|
|
|
|
uint16_t tag; /* command tag */
|
|
|
|
uint16_t class; /* class ID */
|
|
|
|
uint16_t cmd; /* command opcode */
|
|
|
|
};
|
|
|
|
|
|
|
|
|
2005-11-15 20:53:52 +00:00
|
|
|
/* SPU context query/set operations. */
|
|
|
|
struct spu_context_ops {
|
|
|
|
int (*mbox_read) (struct spu_context * ctx, u32 * data);
|
|
|
|
u32(*mbox_stat_read) (struct spu_context * ctx);
|
2005-12-06 03:52:27 +00:00
|
|
|
unsigned int (*mbox_stat_poll)(struct spu_context *ctx,
|
|
|
|
unsigned int events);
|
2005-11-15 20:53:52 +00:00
|
|
|
int (*ibox_read) (struct spu_context * ctx, u32 * data);
|
|
|
|
int (*wbox_write) (struct spu_context * ctx, u32 data);
|
|
|
|
u32(*signal1_read) (struct spu_context * ctx);
|
|
|
|
void (*signal1_write) (struct spu_context * ctx, u32 data);
|
|
|
|
u32(*signal2_read) (struct spu_context * ctx);
|
|
|
|
void (*signal2_write) (struct spu_context * ctx, u32 data);
|
|
|
|
void (*signal1_type_set) (struct spu_context * ctx, u64 val);
|
|
|
|
u64(*signal1_type_get) (struct spu_context * ctx);
|
|
|
|
void (*signal2_type_set) (struct spu_context * ctx, u64 val);
|
|
|
|
u64(*signal2_type_get) (struct spu_context * ctx);
|
|
|
|
u32(*npc_read) (struct spu_context * ctx);
|
|
|
|
void (*npc_write) (struct spu_context * ctx, u32 data);
|
|
|
|
u32(*status_read) (struct spu_context * ctx);
|
|
|
|
char*(*get_ls) (struct spu_context * ctx);
|
2006-11-20 17:45:09 +00:00
|
|
|
u32 (*runcntl_read) (struct spu_context * ctx);
|
2005-12-06 03:52:25 +00:00
|
|
|
void (*runcntl_write) (struct spu_context * ctx, u32 data);
|
2006-11-20 17:45:08 +00:00
|
|
|
void (*master_start) (struct spu_context * ctx);
|
|
|
|
void (*master_stop) (struct spu_context * ctx);
|
2006-03-22 23:00:11 +00:00
|
|
|
int (*set_mfc_query)(struct spu_context * ctx, u32 mask, u32 mode);
|
|
|
|
u32 (*read_mfc_tagstatus)(struct spu_context * ctx);
|
|
|
|
u32 (*get_mfc_free_elements)(struct spu_context *ctx);
|
2006-11-20 17:44:58 +00:00
|
|
|
int (*send_mfc_command)(struct spu_context * ctx,
|
|
|
|
struct mfc_dma_command * cmd);
|
|
|
|
void (*dma_info_read) (struct spu_context * ctx,
|
|
|
|
struct spu_dma_info * info);
|
|
|
|
void (*proxydma_info_read) (struct spu_context * ctx,
|
|
|
|
struct spu_proxydma_info * info);
|
2007-04-23 19:08:15 +00:00
|
|
|
void (*restart_dma)(struct spu_context *ctx);
|
2005-11-15 20:53:48 +00:00
|
|
|
};
|
|
|
|
|
2005-11-15 20:53:52 +00:00
|
|
|
extern struct spu_context_ops spu_hw_ops;
|
|
|
|
extern struct spu_context_ops spu_backing_ops;
|
|
|
|
|
2005-11-15 20:53:48 +00:00
|
|
|
struct spufs_inode_info {
|
|
|
|
struct spu_context *i_ctx;
|
2006-10-04 15:26:15 +00:00
|
|
|
struct spu_gang *i_gang;
|
2005-11-15 20:53:48 +00:00
|
|
|
struct inode vfs_inode;
|
2007-04-23 19:08:07 +00:00
|
|
|
int i_openers;
|
2005-11-15 20:53:48 +00:00
|
|
|
};
|
|
|
|
#define SPUFS_I(inode) \
|
|
|
|
container_of(inode, struct spufs_inode_info, vfs_inode)
|
|
|
|
|
|
|
|
extern struct tree_descr spufs_dir_contents[];
|
2006-10-24 16:31:16 +00:00
|
|
|
extern struct tree_descr spufs_dir_nosched_contents[];
|
2005-11-15 20:53:48 +00:00
|
|
|
|
|
|
|
/* system call implementation */
|
|
|
|
long spufs_run_spu(struct file *file,
|
|
|
|
struct spu_context *ctx, u32 *npc, u32 *status);
|
2006-10-04 15:26:15 +00:00
|
|
|
long spufs_create(struct nameidata *nd,
|
2005-11-15 20:53:48 +00:00
|
|
|
unsigned int flags, mode_t mode);
|
2007-02-12 08:55:37 +00:00
|
|
|
extern const struct file_operations spufs_context_fops;
|
2005-11-15 20:53:48 +00:00
|
|
|
|
2006-10-04 15:26:15 +00:00
|
|
|
/* gang management */
|
|
|
|
struct spu_gang *alloc_spu_gang(void);
|
|
|
|
struct spu_gang *get_spu_gang(struct spu_gang *gang);
|
|
|
|
int put_spu_gang(struct spu_gang *gang);
|
|
|
|
void spu_gang_remove_ctx(struct spu_gang *gang, struct spu_context *ctx);
|
|
|
|
void spu_gang_add_ctx(struct spu_gang *gang, struct spu_context *ctx);
|
|
|
|
|
2007-04-23 19:08:15 +00:00
|
|
|
/* fault handling */
|
|
|
|
int spufs_handle_class1(struct spu_context *ctx);
|
|
|
|
|
2005-11-15 20:53:48 +00:00
|
|
|
/* context management */
|
2007-06-29 00:58:02 +00:00
|
|
|
extern atomic_t nr_spu_contexts;
|
2007-02-13 20:54:21 +00:00
|
|
|
static inline void spu_acquire(struct spu_context *ctx)
|
|
|
|
{
|
|
|
|
mutex_lock(&ctx->state_mutex);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void spu_release(struct spu_context *ctx)
|
|
|
|
{
|
|
|
|
mutex_unlock(&ctx->state_mutex);
|
|
|
|
}
|
|
|
|
|
2006-10-04 15:26:15 +00:00
|
|
|
struct spu_context * alloc_spu_context(struct spu_gang *gang);
|
2005-11-15 20:53:48 +00:00
|
|
|
void destroy_spu_context(struct kref *kref);
|
|
|
|
struct spu_context * get_spu_context(struct spu_context *ctx);
|
|
|
|
int put_spu_context(struct spu_context *ctx);
|
2005-12-06 03:52:25 +00:00
|
|
|
void spu_unmap_mappings(struct spu_context *ctx);
|
2005-11-15 20:53:48 +00:00
|
|
|
|
2005-11-15 20:53:52 +00:00
|
|
|
void spu_forget(struct spu_context *ctx);
|
2007-02-13 20:54:24 +00:00
|
|
|
int spu_acquire_runnable(struct spu_context *ctx, unsigned long flags);
|
2005-11-15 20:53:48 +00:00
|
|
|
void spu_acquire_saved(struct spu_context *ctx);
|
2007-03-09 23:05:36 +00:00
|
|
|
|
2007-02-13 20:54:24 +00:00
|
|
|
int spu_activate(struct spu_context *ctx, unsigned long flags);
|
2005-11-15 20:53:52 +00:00
|
|
|
void spu_deactivate(struct spu_context *ctx);
|
|
|
|
void spu_yield(struct spu_context *ctx);
|
2007-06-29 00:57:52 +00:00
|
|
|
void spu_set_timeslice(struct spu_context *ctx);
|
2007-06-29 00:57:55 +00:00
|
|
|
void spu_update_sched_info(struct spu_context *ctx);
|
|
|
|
void __spu_update_sched_info(struct spu_context *ctx);
|
2005-11-15 20:53:52 +00:00
|
|
|
int __init spu_sched_init(void);
|
|
|
|
void __exit spu_sched_exit(void);
|
|
|
|
|
2006-11-20 17:45:10 +00:00
|
|
|
extern char *isolated_loader;
|
|
|
|
|
2006-01-04 19:31:29 +00:00
|
|
|
/*
|
|
|
|
* spufs_wait
|
2007-06-29 00:58:00 +00:00
|
|
|
* Same as wait_event_interruptible(), except that here
|
2006-01-04 19:31:29 +00:00
|
|
|
* we need to call spu_release(ctx) before sleeping, and
|
|
|
|
* then spu_acquire(ctx) when awoken.
|
|
|
|
*/
|
|
|
|
|
|
|
|
#define spufs_wait(wq, condition) \
|
|
|
|
({ \
|
|
|
|
int __ret = 0; \
|
|
|
|
DEFINE_WAIT(__wait); \
|
|
|
|
for (;;) { \
|
|
|
|
prepare_to_wait(&(wq), &__wait, TASK_INTERRUPTIBLE); \
|
|
|
|
if (condition) \
|
|
|
|
break; \
|
2007-04-23 19:08:24 +00:00
|
|
|
if (signal_pending(current)) { \
|
|
|
|
__ret = -ERESTARTSYS; \
|
|
|
|
break; \
|
2006-01-04 19:31:29 +00:00
|
|
|
} \
|
2007-04-23 19:08:24 +00:00
|
|
|
spu_release(ctx); \
|
|
|
|
schedule(); \
|
|
|
|
spu_acquire(ctx); \
|
2006-01-04 19:31:29 +00:00
|
|
|
} \
|
|
|
|
finish_wait(&(wq), &__wait); \
|
|
|
|
__ret; \
|
|
|
|
})
|
|
|
|
|
2005-11-15 20:53:52 +00:00
|
|
|
size_t spu_wbox_write(struct spu_context *ctx, u32 data);
|
|
|
|
size_t spu_ibox_read(struct spu_context *ctx, u32 *data);
|
|
|
|
|
|
|
|
/* irq callback funcs. */
|
|
|
|
void spufs_ibox_callback(struct spu *spu);
|
|
|
|
void spufs_wbox_callback(struct spu *spu);
|
2005-12-06 03:52:25 +00:00
|
|
|
void spufs_stop_callback(struct spu *spu);
|
2006-03-22 23:00:11 +00:00
|
|
|
void spufs_mfc_callback(struct spu *spu);
|
2006-10-04 15:26:14 +00:00
|
|
|
void spufs_dma_callback(struct spu *spu, int type);
|
2005-11-15 20:53:52 +00:00
|
|
|
|
[POWERPC] coredump: Add SPU elf notes to coredump.
This patch adds SPU elf notes to the coredump. It creates a separate note
for each of /regs, /fpcr, /lslr, /decr, /decr_status, /mem, /signal1,
/signal1_type, /signal2, /signal2_type, /event_mask, /event_status,
/mbox_info, /ibox_info, /wbox_info, /dma_info, /proxydma_info, /object-id.
A new macro, ARCH_HAVE_EXTRA_NOTES, was created for architectures to
specify they have extra elf core notes.
A new macro, ELF_CORE_EXTRA_NOTES_SIZE, was created so the size of the
additional notes could be calculated and added to the notes phdr entry.
A new macro, ELF_CORE_WRITE_EXTRA_NOTES, was created so the new notes
would be written after the existing notes.
The SPU coredump code resides in spufs. Stub functions are provided in the
kernel which are hooked into the spufs code which does the actual work via
register_arch_coredump_calls().
A new set of __spufs_<file>_read/get() functions was provided to allow the
coredump code to read from the spufs files without having to lock the
SPU context for each file read from.
Cc: <linux-arch@vger.kernel.org>
Signed-off-by: Dwayne Grant McConnell <decimal@us.ibm.com>
Signed-off-by: Arnd Bergmann <arnd.bergmann@de.ibm.com>
2006-11-22 23:46:37 +00:00
|
|
|
extern struct spu_coredump_calls spufs_coredump_calls;
|
|
|
|
struct spufs_coredump_reader {
|
|
|
|
char *name;
|
|
|
|
ssize_t (*read)(struct spu_context *ctx,
|
|
|
|
char __user *buffer, size_t size, loff_t *pos);
|
|
|
|
u64 (*get)(void *data);
|
|
|
|
size_t size;
|
|
|
|
};
|
|
|
|
extern struct spufs_coredump_reader spufs_coredump_read[];
|
|
|
|
extern int spufs_coredump_num_notes;
|
|
|
|
|
2007-06-29 00:58:03 +00:00
|
|
|
/*
|
|
|
|
* This function is a little bit too large for an inline, but
|
|
|
|
* as fault.c is built into the kernel we can't move it out of
|
|
|
|
* line.
|
|
|
|
*/
|
|
|
|
static inline void spuctx_switch_state(struct spu_context *ctx,
|
|
|
|
enum spuctx_execution_state new_state)
|
|
|
|
{
|
|
|
|
WARN_ON(!mutex_is_locked(&ctx->state_mutex));
|
|
|
|
|
|
|
|
if (ctx->stats.execution_state != new_state) {
|
|
|
|
unsigned long curtime = jiffies;
|
|
|
|
|
|
|
|
ctx->stats.times[ctx->stats.execution_state] +=
|
|
|
|
curtime - ctx->stats.tstamp;
|
|
|
|
ctx->stats.tstamp = curtime;
|
|
|
|
ctx->stats.execution_state = new_state;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2005-11-15 20:53:48 +00:00
|
|
|
#endif
|