/* * Byte threshold to limit memory consumption for flip buffers. * The actual memory limit is > 2x this amount.
*/ #define TTYB_DEFAULT_MEM_LIMIT (640 * 1024UL)
/* * We default to dicing tty buffer allocations to this many characters * in order to avoid multiple page allocations. We know the size of * tty_buffer itself but it must also be taken into account that the * buffer is 256 byte aligned. See tty_buffer_find for the allocation * logic this must match.
*/
/** * tty_buffer_lock_exclusive - gain exclusive access to buffer * @port: tty port owning the flip buffer * * Guarantees safe use of the &tty_ldisc_ops.receive_buf() method by excluding * the buffer work and any pending flush from using the flip buffer. Data can * continue to be added concurrently to the flip buffer from the driver side. * * See also tty_buffer_unlock_exclusive().
*/ void tty_buffer_lock_exclusive(struct tty_port *port)
{ struct tty_bufhead *buf = &port->buf;
/** * tty_buffer_unlock_exclusive - release exclusive access * @port: tty port owning the flip buffer * * The buffer work is restarted if there is data in the flip buffer. * * See also tty_buffer_lock_exclusive().
*/ void tty_buffer_unlock_exclusive(struct tty_port *port)
{ struct tty_bufhead *buf = &port->buf; bool restart = buf->head->commit != buf->head->read;
if (restart)
queue_work(system_unbound_wq, &buf->work);
}
EXPORT_SYMBOL_GPL(tty_buffer_unlock_exclusive);
/** * tty_buffer_space_avail - return unused buffer space * @port: tty port owning the flip buffer * * Returns: the # of bytes which can be written by the driver without reaching * the buffer limit. * * Note: this does not guarantee that memory is available to write the returned * # of bytes (use tty_prepare_flip_string() to pre-allocate if memory * guarantee is required).
*/ unsignedint tty_buffer_space_avail(struct tty_port *port)
{ int space = port->buf.mem_limit - atomic_read(&port->buf.mem_used);
/** * tty_buffer_free_all - free buffers used by a tty * @port: tty port to free from * * Remove all the buffers pending on a tty whether queued with data or in the * free ring. Must be called when the tty is no longer in use.
*/ void tty_buffer_free_all(struct tty_port *port)
{ struct tty_bufhead *buf = &port->buf; struct tty_buffer *p, *next; struct llist_node *llist; unsignedint freed = 0; int still_used;
still_used = atomic_xchg(&buf->mem_used, 0);
WARN(still_used != freed, "we still have not freed %d bytes!",
still_used - freed);
}
/** * tty_buffer_alloc - allocate a tty buffer * @port: tty port * @size: desired size (characters) * * Allocate a new tty buffer to hold the desired number of characters. We * round our buffers off in 256 character chunks to get better allocation * behaviour. * * Returns: %NULL if out of memory or the allocation would exceed the per * device queue.
*/ staticstruct tty_buffer *tty_buffer_alloc(struct tty_port *port, size_t size)
{ struct llist_node *free; struct tty_buffer *p;
/* Round the buffer size out */
size = __ALIGN_MASK(size, TTYB_ALIGN_MASK);
if (size <= MIN_TTYB_SIZE) {
free = llist_del_first(&port->buf.free); if (free) {
p = llist_entry(free, struct tty_buffer, free); goto found;
}
}
/* Should possibly check if this fails for the largest buffer we * have queued and recycle that ?
*/ if (atomic_read(&port->buf.mem_used) > port->buf.mem_limit) return NULL;
p = kmalloc(struct_size(p, data, 2 * size), GFP_ATOMIC | __GFP_NOWARN); if (p == NULL) return NULL;
/** * tty_buffer_free - free a tty buffer * @port: tty port owning the buffer * @b: the buffer to free * * Free a tty buffer, or add it to the free list according to our internal * strategy.
*/ staticvoid tty_buffer_free(struct tty_port *port, struct tty_buffer *b)
{ struct tty_bufhead *buf = &port->buf;
/* Dumb strategy for now - should keep some stats */
WARN_ON(atomic_sub_return(b->size, &buf->mem_used) < 0);
/** * __tty_buffer_request_room - grow tty buffer if needed * @port: tty port * @size: size desired * @flags: buffer has to store flags along character data * * Make at least @size bytes of linear space available for the tty buffer. * * Will change over to a new buffer if the current buffer is encoded as * %TTY_NORMAL (so has no flags buffer) and the new buffer requires a flags * buffer. * * Returns: the size we managed to find.
*/ staticint __tty_buffer_request_room(struct tty_port *port, size_t size, bool flags)
{ struct tty_bufhead *buf = &port->buf; struct tty_buffer *n, *b = buf->tail;
size_t left = (b->flags ? 1 : 2) * b->size - b->used; bool change = !b->flags && flags;
if (!change && left >= size) return size;
/* This is the slow path - looking for new buffers to use */
n = tty_buffer_alloc(port, size); if (n == NULL) return change ? 0 : left;
n->flags = flags;
buf->tail = n; /* * Paired w/ acquire in flush_to_ldisc() and lookahead_bufs() * ensures they see all buffer data.
*/
smp_store_release(&b->commit, b->used); /* * Paired w/ acquire in flush_to_ldisc() and lookahead_bufs() * ensures the latest commit value can be read before the head * is advanced to the next buffer.
*/
smp_store_release(&b->next, n);
/* There is a small chance that we need to split the data over * several buffers. If this is the case we must loop.
*/
} while (unlikely(size > copied));
/** * tty_prepare_flip_string - make room for characters * @port: tty port * @chars: return pointer for character write area * @size: desired size * * Prepare a block of space in the buffer for data. * * This is used for drivers that need their own block copy routines into the * buffer. There is no guarantee the buffer is a DMA target! * * Returns: the length available and buffer pointer (@chars) to the space which * is now allocated and accounted for as ready for normal characters.
*/
size_t tty_prepare_flip_string(struct tty_port *port, u8 **chars, size_t size)
{
size_t space = __tty_buffer_request_room(port, size, false);
if (likely(space)) { struct tty_buffer *tb = port->buf.tail;
/** * tty_ldisc_receive_buf - forward data to line discipline * @ld: line discipline to process input * @p: char buffer * @f: %TTY_NORMAL, %TTY_BREAK, etc. flags buffer * @count: number of bytes to process * * Callers other than flush_to_ldisc() need to exclude the kworker from * concurrent use of the line discipline, see paste_selection(). * * Returns: the number of bytes processed.
*/
size_t tty_ldisc_receive_buf(struct tty_ldisc *ld, const u8 *p, const u8 *f,
size_t count)
{ if (ld->ops->receive_buf2)
count = ld->ops->receive_buf2(ld->tty, p, f, count); else {
count = min_t(size_t, count, ld->tty->receive_room); if (count && ld->ops->receive_buf)
ld->ops->receive_buf(ld->tty, p, f, count);
} return count;
}
EXPORT_SYMBOL_GPL(tty_ldisc_receive_buf);
while (head) { struct tty_buffer *next; unsignedint count;
/* * Paired w/ release in __tty_buffer_request_room(); * ensures commit value read is not stale if the head * is advancing to the next buffer.
*/
next = smp_load_acquire(&head->next); /* * Paired w/ release in __tty_buffer_request_room() or in * tty_buffer_flush(); ensures we see the committed buffer data.
*/
count = smp_load_acquire(&head->commit) - head->lookahead; if (!count) {
head = next; continue;
}
if (port->client_ops->lookahead_buf) {
u8 *p, *f = NULL;
p = char_buf_ptr(head, head->lookahead); if (head->flags)
f = flag_buf_ptr(head, head->lookahead);
if (head->flags)
f = flag_buf_ptr(head, head->read);
n = port->client_ops->receive_buf(port, p, f, count); if (n > 0)
memset(p, 0, n); return n;
}
/** * flush_to_ldisc - flush data from buffer to ldisc * @work: tty structure passed from work queue. * * This routine is called out of the software interrupt to flush data from the * buffer chain to the line discipline. * * The receive_buf() method is single threaded for each tty instance. * * Locking: takes buffer lock to ensure single-threaded flip buffer 'consumer'.
*/ staticvoid flush_to_ldisc(struct work_struct *work)
{ struct tty_port *port = container_of(work, struct tty_port, buf.work); struct tty_bufhead *buf = &port->buf;
/* Ldisc or user is trying to gain exclusive access */ if (atomic_read(&buf->priority)) break;
/* paired w/ release in __tty_buffer_request_room(); * ensures commit value read is not stale if the head * is advancing to the next buffer
*/
next = smp_load_acquire(&head->next); /* paired w/ release in __tty_buffer_request_room() or in * tty_buffer_flush(); ensures we see the committed buffer data
*/
count = smp_load_acquire(&head->commit) - head->read; if (!count) { if (next == NULL) break;
buf->head = next;
tty_buffer_free(port, head); continue;
}
rcvd = receive_buf(port, head, count);
head->read += rcvd; if (rcvd < count)
lookahead_bufs(port, head); if (!rcvd) break;
/** * tty_flip_buffer_push - push terminal buffers * @port: tty port to push * * Queue a push of the terminal flip buffers to the line discipline. Can be * called from IRQ/atomic context. * * In the event of the queue being busy for flipping the work will be held off * and retried later.
*/ void tty_flip_buffer_push(struct tty_port *port)
{ struct tty_bufhead *buf = &port->buf;
/** * tty_insert_flip_string_and_push_buffer - add characters to the tty buffer and * push * @port: tty port * @chars: characters * @size: size * * The function combines tty_insert_flip_string() and tty_flip_buffer_push() * with the exception of properly holding the @port->lock. * * To be used only internally (by pty currently). * * Returns: the number added.
*/ int tty_insert_flip_string_and_push_buffer(struct tty_port *port, const u8 *chars, size_t size)
{ struct tty_bufhead *buf = &port->buf; unsignedlong flags;
/** * tty_buffer_init - prepare a tty buffer structure * @port: tty port to initialise * * Set up the initial state of the buffer management for a tty device. Must be * called before the other tty buffer functions are used.
*/ void tty_buffer_init(struct tty_port *port)
{ struct tty_bufhead *buf = &port->buf;
/** * tty_buffer_set_limit - change the tty buffer memory limit * @port: tty port to change * @limit: memory limit to set * * Change the tty buffer memory limit. * * Must be called before the other tty buffer functions are used.
*/ int tty_buffer_set_limit(struct tty_port *port, int limit)
{ if (limit < MIN_TTYB_SIZE) return -EINVAL;
port->buf.mem_limit = limit; return 0;
}
EXPORT_SYMBOL_GPL(tty_buffer_set_limit);
/* slave ptys can claim nested buffer lock when handling BRK and INTR */ void tty_buffer_set_lock_subclass(struct tty_port *port)
{
lockdep_set_subclass(&port->buf.lock, TTY_LOCK_SLAVE);
}
Die Informationen auf dieser Webseite wurden
nach bestem Wissen sorgfältig zusammengestellt. Es wird jedoch weder Vollständigkeit, noch Richtigkeit,
noch Qualität der bereit gestellten Informationen zugesichert.
Bemerkung:
Die farbliche Syntaxdarstellung und die Messung sind noch experimentell.